Dec 09 16:56:42 crc systemd[1]: Starting Kubernetes Kubelet... Dec 09 16:56:42 crc restorecon[4732]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:42 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:43 crc restorecon[4732]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 16:56:43 crc restorecon[4732]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 09 16:56:43 crc kubenswrapper[4954]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 16:56:43 crc kubenswrapper[4954]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 09 16:56:43 crc kubenswrapper[4954]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 16:56:43 crc kubenswrapper[4954]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 16:56:43 crc kubenswrapper[4954]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 09 16:56:43 crc kubenswrapper[4954]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.932549 4954 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.938927 4954 feature_gate.go:330] unrecognized feature gate: Example Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.938965 4954 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.938976 4954 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.938985 4954 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.938994 4954 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939003 4954 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939014 4954 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939027 4954 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939038 4954 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939047 4954 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939056 4954 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939064 4954 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939073 4954 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939083 4954 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939092 4954 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939101 4954 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939111 4954 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939119 4954 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939201 4954 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939744 4954 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939795 4954 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939800 4954 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939805 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939809 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939813 4954 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939818 4954 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939822 4954 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939826 4954 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939830 4954 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939835 4954 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939838 4954 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939842 4954 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939853 4954 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939858 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939862 4954 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939866 4954 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939870 4954 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939874 4954 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939878 4954 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939882 4954 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939890 4954 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939897 4954 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939903 4954 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939911 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939916 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939919 4954 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939924 4954 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939928 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939934 4954 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939941 4954 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939947 4954 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939952 4954 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939957 4954 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939960 4954 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939964 4954 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939970 4954 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939977 4954 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939982 4954 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939986 4954 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939990 4954 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939994 4954 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.939997 4954 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.940001 4954 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.940004 4954 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.940010 4954 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.940014 4954 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.940018 4954 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.940024 4954 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.940029 4954 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.940034 4954 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.940038 4954 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940790 4954 flags.go:64] FLAG: --address="0.0.0.0" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940851 4954 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940869 4954 flags.go:64] FLAG: --anonymous-auth="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940882 4954 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940896 4954 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940939 4954 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940952 4954 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940964 4954 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940974 4954 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940984 4954 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.940996 4954 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941006 4954 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941016 4954 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941032 4954 flags.go:64] FLAG: --cgroup-root="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941042 4954 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941052 4954 flags.go:64] FLAG: --client-ca-file="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941063 4954 flags.go:64] FLAG: --cloud-config="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941072 4954 flags.go:64] FLAG: --cloud-provider="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941082 4954 flags.go:64] FLAG: --cluster-dns="[]" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941093 4954 flags.go:64] FLAG: --cluster-domain="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941103 4954 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941114 4954 flags.go:64] FLAG: --config-dir="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941129 4954 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941141 4954 flags.go:64] FLAG: --container-log-max-files="5" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941156 4954 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941166 4954 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941177 4954 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941188 4954 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941198 4954 flags.go:64] FLAG: --contention-profiling="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941207 4954 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941222 4954 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941233 4954 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941244 4954 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941257 4954 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941267 4954 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941277 4954 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941336 4954 flags.go:64] FLAG: --enable-load-reader="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941652 4954 flags.go:64] FLAG: --enable-server="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941672 4954 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941688 4954 flags.go:64] FLAG: --event-burst="100" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941699 4954 flags.go:64] FLAG: --event-qps="50" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941709 4954 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941719 4954 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941729 4954 flags.go:64] FLAG: --eviction-hard="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941742 4954 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941751 4954 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941761 4954 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941772 4954 flags.go:64] FLAG: --eviction-soft="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941784 4954 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941797 4954 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941806 4954 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941816 4954 flags.go:64] FLAG: --experimental-mounter-path="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941826 4954 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941836 4954 flags.go:64] FLAG: --fail-swap-on="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941847 4954 flags.go:64] FLAG: --feature-gates="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941859 4954 flags.go:64] FLAG: --file-check-frequency="20s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941869 4954 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941880 4954 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941890 4954 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941900 4954 flags.go:64] FLAG: --healthz-port="10248" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941911 4954 flags.go:64] FLAG: --help="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941921 4954 flags.go:64] FLAG: --hostname-override="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941931 4954 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941941 4954 flags.go:64] FLAG: --http-check-frequency="20s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941950 4954 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941963 4954 flags.go:64] FLAG: --image-credential-provider-config="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941972 4954 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941982 4954 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.941992 4954 flags.go:64] FLAG: --image-service-endpoint="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942001 4954 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942010 4954 flags.go:64] FLAG: --kube-api-burst="100" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942020 4954 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942031 4954 flags.go:64] FLAG: --kube-api-qps="50" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942040 4954 flags.go:64] FLAG: --kube-reserved="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942050 4954 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942059 4954 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942069 4954 flags.go:64] FLAG: --kubelet-cgroups="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942078 4954 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942089 4954 flags.go:64] FLAG: --lock-file="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942098 4954 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942107 4954 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942117 4954 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942131 4954 flags.go:64] FLAG: --log-json-split-stream="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942141 4954 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942150 4954 flags.go:64] FLAG: --log-text-split-stream="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942160 4954 flags.go:64] FLAG: --logging-format="text" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942169 4954 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942179 4954 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942190 4954 flags.go:64] FLAG: --manifest-url="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942199 4954 flags.go:64] FLAG: --manifest-url-header="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942212 4954 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942222 4954 flags.go:64] FLAG: --max-open-files="1000000" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942234 4954 flags.go:64] FLAG: --max-pods="110" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942243 4954 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942254 4954 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942264 4954 flags.go:64] FLAG: --memory-manager-policy="None" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942273 4954 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942284 4954 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942294 4954 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942305 4954 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942330 4954 flags.go:64] FLAG: --node-status-max-images="50" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942340 4954 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942350 4954 flags.go:64] FLAG: --oom-score-adj="-999" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942360 4954 flags.go:64] FLAG: --pod-cidr="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942370 4954 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942387 4954 flags.go:64] FLAG: --pod-manifest-path="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942397 4954 flags.go:64] FLAG: --pod-max-pids="-1" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942407 4954 flags.go:64] FLAG: --pods-per-core="0" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942448 4954 flags.go:64] FLAG: --port="10250" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942458 4954 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942468 4954 flags.go:64] FLAG: --provider-id="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942477 4954 flags.go:64] FLAG: --qos-reserved="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942487 4954 flags.go:64] FLAG: --read-only-port="10255" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942496 4954 flags.go:64] FLAG: --register-node="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942506 4954 flags.go:64] FLAG: --register-schedulable="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942516 4954 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942532 4954 flags.go:64] FLAG: --registry-burst="10" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942543 4954 flags.go:64] FLAG: --registry-qps="5" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942552 4954 flags.go:64] FLAG: --reserved-cpus="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942562 4954 flags.go:64] FLAG: --reserved-memory="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942574 4954 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942584 4954 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942620 4954 flags.go:64] FLAG: --rotate-certificates="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942630 4954 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942640 4954 flags.go:64] FLAG: --runonce="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942650 4954 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942660 4954 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942670 4954 flags.go:64] FLAG: --seccomp-default="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942679 4954 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942689 4954 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942701 4954 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942711 4954 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942721 4954 flags.go:64] FLAG: --storage-driver-password="root" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942731 4954 flags.go:64] FLAG: --storage-driver-secure="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942740 4954 flags.go:64] FLAG: --storage-driver-table="stats" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942750 4954 flags.go:64] FLAG: --storage-driver-user="root" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942760 4954 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942770 4954 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942780 4954 flags.go:64] FLAG: --system-cgroups="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942789 4954 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942807 4954 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942819 4954 flags.go:64] FLAG: --tls-cert-file="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942829 4954 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942842 4954 flags.go:64] FLAG: --tls-min-version="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942851 4954 flags.go:64] FLAG: --tls-private-key-file="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942861 4954 flags.go:64] FLAG: --topology-manager-policy="none" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942871 4954 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942880 4954 flags.go:64] FLAG: --topology-manager-scope="container" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942890 4954 flags.go:64] FLAG: --v="2" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942903 4954 flags.go:64] FLAG: --version="false" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942916 4954 flags.go:64] FLAG: --vmodule="" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942927 4954 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.942937 4954 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943159 4954 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943169 4954 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943178 4954 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943187 4954 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943196 4954 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943205 4954 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943214 4954 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943222 4954 feature_gate.go:330] unrecognized feature gate: Example Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943231 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943240 4954 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943248 4954 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943257 4954 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943268 4954 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943280 4954 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943290 4954 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943300 4954 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943308 4954 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943316 4954 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943325 4954 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943333 4954 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943341 4954 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943350 4954 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943358 4954 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943368 4954 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943377 4954 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943385 4954 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943393 4954 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943401 4954 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943410 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943418 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943427 4954 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943435 4954 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943443 4954 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943451 4954 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943460 4954 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943468 4954 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943476 4954 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943487 4954 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943498 4954 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943509 4954 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943521 4954 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943531 4954 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943542 4954 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943551 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943559 4954 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943568 4954 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943577 4954 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943585 4954 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943617 4954 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943626 4954 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943635 4954 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943645 4954 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943653 4954 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943661 4954 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943669 4954 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943678 4954 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943688 4954 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943696 4954 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943705 4954 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943713 4954 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943722 4954 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943730 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943741 4954 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943750 4954 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943759 4954 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943768 4954 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943776 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943785 4954 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943793 4954 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943801 4954 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.943810 4954 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.944170 4954 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.953443 4954 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.953498 4954 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953629 4954 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953649 4954 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953661 4954 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953670 4954 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953676 4954 feature_gate.go:330] unrecognized feature gate: Example Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953682 4954 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953687 4954 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953693 4954 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953698 4954 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953703 4954 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953709 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953714 4954 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953719 4954 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953724 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953730 4954 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953735 4954 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953740 4954 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953745 4954 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953750 4954 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953756 4954 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953761 4954 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953766 4954 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953771 4954 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953778 4954 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953787 4954 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953795 4954 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953801 4954 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953807 4954 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953813 4954 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953819 4954 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953825 4954 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953830 4954 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953838 4954 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953844 4954 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953853 4954 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953859 4954 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953865 4954 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953871 4954 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953876 4954 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953881 4954 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953888 4954 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953893 4954 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953899 4954 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953904 4954 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953909 4954 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953915 4954 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953920 4954 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953926 4954 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953933 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953938 4954 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953943 4954 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953948 4954 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953954 4954 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953959 4954 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953964 4954 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953969 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953977 4954 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953982 4954 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953988 4954 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.953994 4954 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954001 4954 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954007 4954 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954012 4954 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954017 4954 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954023 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954028 4954 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954033 4954 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954038 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954043 4954 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954049 4954 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954055 4954 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.954064 4954 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954213 4954 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954223 4954 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954229 4954 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954235 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954240 4954 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954245 4954 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954250 4954 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954256 4954 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954261 4954 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954266 4954 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954271 4954 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954277 4954 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954282 4954 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954287 4954 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954292 4954 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954297 4954 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954303 4954 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954309 4954 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954316 4954 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954322 4954 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954328 4954 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954334 4954 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954340 4954 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954345 4954 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954350 4954 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954355 4954 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954360 4954 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954365 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954371 4954 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954377 4954 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954385 4954 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954428 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954434 4954 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954440 4954 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954446 4954 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954451 4954 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954457 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954462 4954 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954468 4954 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954473 4954 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954478 4954 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954483 4954 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954488 4954 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954494 4954 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954499 4954 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954504 4954 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954509 4954 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954514 4954 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954519 4954 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954525 4954 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954531 4954 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954536 4954 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954541 4954 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954546 4954 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954551 4954 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954557 4954 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954562 4954 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954567 4954 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954572 4954 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954577 4954 feature_gate.go:330] unrecognized feature gate: Example Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954582 4954 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954588 4954 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954616 4954 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954624 4954 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954631 4954 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954638 4954 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954643 4954 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954648 4954 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954654 4954 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954659 4954 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 16:56:43 crc kubenswrapper[4954]: W1209 16:56:43.954668 4954 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.954677 4954 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.954921 4954 server.go:940] "Client rotation is on, will bootstrap in background" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.959319 4954 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.959474 4954 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.960257 4954 server.go:997] "Starting client certificate rotation" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.960299 4954 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.960501 4954 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-08 04:54:00.202293335 +0000 UTC Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.960640 4954 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 707h57m16.241661131s for next certificate rotation Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.974209 4954 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.977100 4954 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 16:56:43 crc kubenswrapper[4954]: I1209 16:56:43.989346 4954 log.go:25] "Validated CRI v1 runtime API" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.015434 4954 log.go:25] "Validated CRI v1 image API" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.017260 4954 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.020345 4954 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-09-16-51-58-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.020382 4954 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.045987 4954 manager.go:217] Machine: {Timestamp:2025-12-09 16:56:44.04187818 +0000 UTC m=+0.430052080 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:d423a69c-3c1d-438f-aa6d-1e5bcc3d983e BootID:936ce91d-f269-4212-970c-75535102a27e Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:03:5f:48 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:03:5f:48 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:cb:f1:9b Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5f:60:ef Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:30:f9:03 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:7e:5b:81 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:9a:aa:77 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:5a:85:3c:6a:c5:e0 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9a:cd:4d:3b:b4:e1 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.046450 4954 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.046835 4954 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.047457 4954 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.047645 4954 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.047674 4954 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.047859 4954 topology_manager.go:138] "Creating topology manager with none policy" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.047869 4954 container_manager_linux.go:303] "Creating device plugin manager" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.048090 4954 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.048118 4954 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.048282 4954 state_mem.go:36] "Initialized new in-memory state store" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.048654 4954 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.049629 4954 kubelet.go:418] "Attempting to sync node with API server" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.049649 4954 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.049671 4954 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.049683 4954 kubelet.go:324] "Adding apiserver pod source" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.049693 4954 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.051353 4954 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.051944 4954 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.053646 4954 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.054523 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.054822 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.054952 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.055064 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.055178 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.055274 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.055383 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.055501 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.055633 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.055756 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.055862 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.055985 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 09 16:56:44 crc kubenswrapper[4954]: W1209 16:56:44.054850 4954 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.056249 4954 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 09 16:56:44 crc kubenswrapper[4954]: W1209 16:56:44.054947 4954 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.056321 4954 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.056964 4954 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.057672 4954 server.go:1280] "Started kubelet" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.058026 4954 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.058143 4954 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.058970 4954 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:44 crc systemd[1]: Started Kubernetes Kubelet. Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.060079 4954 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.062881 4954 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.062958 4954 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 03:16:07.659658529 +0000 UTC Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.063026 4954 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.063698 4954 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.063725 4954 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.063957 4954 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.064888 4954 server.go:460] "Adding debug handlers to kubelet server" Dec 09 16:56:44 crc kubenswrapper[4954]: W1209 16:56:44.071968 4954 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.072067 4954 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.072871 4954 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.073375 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.073000 4954 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187f9a747d106ed8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 16:56:44.057636568 +0000 UTC m=+0.445810418,LastTimestamp:2025-12-09 16:56:44.057636568 +0000 UTC m=+0.445810418,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.074156 4954 factory.go:153] Registering CRI-O factory Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.074280 4954 factory.go:221] Registration of the crio container factory successfully Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.074358 4954 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.074368 4954 factory.go:55] Registering systemd factory Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.074377 4954 factory.go:221] Registration of the systemd container factory successfully Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.074397 4954 factory.go:103] Registering Raw factory Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.074415 4954 manager.go:1196] Started watching for new ooms in manager Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.075185 4954 manager.go:319] Starting recovery of all containers Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078471 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078530 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078546 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078558 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078570 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078581 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078611 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078623 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078635 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078647 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078658 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078672 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078683 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078696 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078708 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078719 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078733 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078744 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078755 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078765 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078776 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078788 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078798 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078808 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078819 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078832 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078847 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078869 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078881 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078892 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078903 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078915 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078928 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078939 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078951 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078963 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078975 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078986 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.078999 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079012 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079026 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079038 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079052 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079064 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079075 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079086 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079099 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079112 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079127 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079138 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079149 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079161 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079177 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079190 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079202 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079216 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079227 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079238 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079250 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079261 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079272 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079287 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079299 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079311 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079347 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079359 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079370 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079387 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079399 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079409 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079419 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079430 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079442 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079453 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079464 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079477 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079488 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079520 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079532 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079543 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079556 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079567 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079579 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079605 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079617 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079630 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079642 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079654 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079665 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079677 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079688 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079700 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079713 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079728 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079739 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079750 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079763 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079774 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079786 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079801 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079814 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079826 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079840 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079854 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079874 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079887 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079900 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079912 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079923 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079935 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079949 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079962 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079975 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.079986 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080000 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080011 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080024 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080036 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080056 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080067 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080078 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080088 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080098 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080110 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080124 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080138 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080148 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080159 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080170 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080181 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080191 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080208 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080218 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080229 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080241 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080252 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080261 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080272 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080285 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080296 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080307 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080317 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080329 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080340 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080351 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080361 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080373 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080385 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080396 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080407 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080418 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080429 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080441 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080464 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080477 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080488 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080501 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080515 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080526 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080536 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080549 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080560 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080572 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080586 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080613 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080625 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080635 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080647 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080656 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080667 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080677 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080687 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080698 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080708 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080717 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080727 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080739 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080748 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080757 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080767 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080777 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080787 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080797 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080808 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080818 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080828 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080839 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080850 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080862 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080876 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080888 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080905 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080917 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080927 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080937 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080952 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080963 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080973 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.080985 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081642 4954 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081664 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081677 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081689 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081701 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081711 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081722 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081732 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081743 4954 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081755 4954 reconstruct.go:97] "Volume reconstruction finished" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.081763 4954 reconciler.go:26] "Reconciler: start to sync state" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.099933 4954 manager.go:324] Recovery completed Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.108107 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.109400 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.109443 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.109456 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.110326 4954 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.110342 4954 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.110360 4954 state_mem.go:36] "Initialized new in-memory state store" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.117224 4954 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.118859 4954 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.118891 4954 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.118917 4954 kubelet.go:2335] "Starting kubelet main sync loop" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.118956 4954 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 09 16:56:44 crc kubenswrapper[4954]: W1209 16:56:44.170095 4954 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.170184 4954 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.173554 4954 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.175491 4954 policy_none.go:49] "None policy: Start" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.176294 4954 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.176319 4954 state_mem.go:35] "Initializing new in-memory state store" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.219644 4954 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.229743 4954 manager.go:334] "Starting Device Plugin manager" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.229817 4954 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.229831 4954 server.go:79] "Starting device plugin registration server" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.230299 4954 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.230319 4954 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.230533 4954 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.230636 4954 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.230647 4954 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.237732 4954 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.275574 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.331182 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.332653 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.332709 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.332730 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.332771 4954 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.333377 4954 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.420292 4954 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.420435 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.422512 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.422583 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.422640 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.422841 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.423115 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.423219 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.424518 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.424575 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.424629 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.424904 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.425088 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.425167 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.425351 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.425453 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.425521 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.426105 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.426147 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.426156 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.426295 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.426326 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.426370 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.426393 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.426554 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.426654 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.427458 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.427515 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.427537 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.427764 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.427917 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.427966 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.428133 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.428187 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.428210 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.428811 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.428867 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.428921 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.429041 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.429072 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.429085 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.429288 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.429332 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.430637 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.430683 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.430708 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485286 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485331 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485358 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485379 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485401 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485421 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485441 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485490 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485572 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485706 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485776 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.485825 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.486406 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.486689 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.486756 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.533788 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.535383 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.535433 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.535455 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.535494 4954 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.536094 4954 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588002 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588063 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588087 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588109 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588135 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588150 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588164 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588182 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588206 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588224 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588246 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588259 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588272 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588284 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588299 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588720 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588779 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588776 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588803 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588829 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588834 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588851 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588868 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588872 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588888 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588906 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588911 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588926 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588947 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.588965 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.677095 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.767676 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.779990 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.793411 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: W1209 16:56:44.811810 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-bec9caa7029f8b22392706e7ab4aa38f31b0eb36403a4aee4bb973c0829d050e WatchSource:0}: Error finding container bec9caa7029f8b22392706e7ab4aa38f31b0eb36403a4aee4bb973c0829d050e: Status 404 returned error can't find the container with id bec9caa7029f8b22392706e7ab4aa38f31b0eb36403a4aee4bb973c0829d050e Dec 09 16:56:44 crc kubenswrapper[4954]: W1209 16:56:44.814746 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-24a2d5b1b1fb64869881069c2a94c1cc994dcb00619dc65d1a8d7687dc7e5a4b WatchSource:0}: Error finding container 24a2d5b1b1fb64869881069c2a94c1cc994dcb00619dc65d1a8d7687dc7e5a4b: Status 404 returned error can't find the container with id 24a2d5b1b1fb64869881069c2a94c1cc994dcb00619dc65d1a8d7687dc7e5a4b Dec 09 16:56:44 crc kubenswrapper[4954]: W1209 16:56:44.823337 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-587b0b4436928c9b5c53dbb7c401cb6026cbc88d61050f68dd0d2f5f614d9bfa WatchSource:0}: Error finding container 587b0b4436928c9b5c53dbb7c401cb6026cbc88d61050f68dd0d2f5f614d9bfa: Status 404 returned error can't find the container with id 587b0b4436928c9b5c53dbb7c401cb6026cbc88d61050f68dd0d2f5f614d9bfa Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.834910 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.846434 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 16:56:44 crc kubenswrapper[4954]: W1209 16:56:44.856427 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-eddaddac0896771339a7c0317a59525f64be3aa90093b8b3ebf320e09119c180 WatchSource:0}: Error finding container eddaddac0896771339a7c0317a59525f64be3aa90093b8b3ebf320e09119c180: Status 404 returned error can't find the container with id eddaddac0896771339a7c0317a59525f64be3aa90093b8b3ebf320e09119c180 Dec 09 16:56:44 crc kubenswrapper[4954]: W1209 16:56:44.858868 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-49f844c1a3fdd71bbe3349f97712c6578b97ebb0d1ea584b264233a74a30ef5c WatchSource:0}: Error finding container 49f844c1a3fdd71bbe3349f97712c6578b97ebb0d1ea584b264233a74a30ef5c: Status 404 returned error can't find the container with id 49f844c1a3fdd71bbe3349f97712c6578b97ebb0d1ea584b264233a74a30ef5c Dec 09 16:56:44 crc kubenswrapper[4954]: W1209 16:56:44.889340 4954 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.889460 4954 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.936433 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.937466 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.937504 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.937518 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:44 crc kubenswrapper[4954]: I1209 16:56:44.937541 4954 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 16:56:44 crc kubenswrapper[4954]: E1209 16:56:44.937877 4954 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Dec 09 16:56:45 crc kubenswrapper[4954]: W1209 16:56:45.036900 4954 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:45 crc kubenswrapper[4954]: E1209 16:56:45.037030 4954 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.061216 4954 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.063302 4954 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 03:22:41.805078145 +0000 UTC Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.063350 4954 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 10h25m56.741732883s for next certificate rotation Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.122898 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"49f844c1a3fdd71bbe3349f97712c6578b97ebb0d1ea584b264233a74a30ef5c"} Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.123845 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"eddaddac0896771339a7c0317a59525f64be3aa90093b8b3ebf320e09119c180"} Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.124649 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"587b0b4436928c9b5c53dbb7c401cb6026cbc88d61050f68dd0d2f5f614d9bfa"} Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.125532 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"24a2d5b1b1fb64869881069c2a94c1cc994dcb00619dc65d1a8d7687dc7e5a4b"} Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.126307 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"bec9caa7029f8b22392706e7ab4aa38f31b0eb36403a4aee4bb973c0829d050e"} Dec 09 16:56:45 crc kubenswrapper[4954]: W1209 16:56:45.469135 4954 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:45 crc kubenswrapper[4954]: E1209 16:56:45.469508 4954 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 09 16:56:45 crc kubenswrapper[4954]: E1209 16:56:45.478961 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Dec 09 16:56:45 crc kubenswrapper[4954]: W1209 16:56:45.643660 4954 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:45 crc kubenswrapper[4954]: E1209 16:56:45.643776 4954 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.738145 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.739826 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.739880 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.739892 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:45 crc kubenswrapper[4954]: I1209 16:56:45.739920 4954 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 16:56:45 crc kubenswrapper[4954]: E1209 16:56:45.740431 4954 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.061231 4954 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.131695 4954 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf" exitCode=0 Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.131757 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf"} Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.131936 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.133271 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.133316 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.133330 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.134816 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.134813 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17"} Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.134955 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909"} Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.134978 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633"} Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.135000 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f"} Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.135724 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.135764 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.135778 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.137538 4954 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef" exitCode=0 Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.137647 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef"} Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.137731 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.138797 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.138831 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.138844 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.139383 4954 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9" exitCode=0 Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.139462 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9"} Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.139475 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.140310 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.140334 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.140344 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.142000 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.142410 4954 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82" exitCode=0 Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.142475 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82"} Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.142569 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.142951 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.142997 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.143013 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.143780 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.143816 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:46 crc kubenswrapper[4954]: I1209 16:56:46.143832 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.060710 4954 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Dec 09 16:56:47 crc kubenswrapper[4954]: E1209 16:56:47.079542 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="3.2s" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.087821 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.147112 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b"} Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.147153 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0"} Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.147163 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7"} Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.147171 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb"} Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.147179 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62"} Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.147192 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.148171 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.148203 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.148214 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.148901 4954 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b" exitCode=0 Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.148956 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b"} Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.149046 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.150005 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.150033 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.150046 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.152538 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"17199e1bc8be0a9cb3e83dca2669f2d58187f64c1b82d3bf4a84b1a39538ec03"} Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.152637 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.153307 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.153333 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.153346 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.155451 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.155870 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.156189 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"23844430c0b5f37c59d048d297bf804c4fbc96cddc4958465f909b24e2635e51"} Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.156216 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e462de97284638e1161f14f470c25bad53981d0d08f81dde5980991e189edecc"} Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.156230 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"327ec4b9b835a025b2efad8b30d5eab591d0860da7b3a214553b94b9b026c7f2"} Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.156662 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.156689 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.156700 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.157189 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.157212 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.157223 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.285412 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.332736 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.340930 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.342066 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.342093 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.342103 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.342123 4954 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.553562 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:47 crc kubenswrapper[4954]: I1209 16:56:47.575792 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.160347 4954 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63" exitCode=0 Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.160455 4954 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.160485 4954 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.160495 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.160515 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.160517 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161006 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161278 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161295 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63"} Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161682 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161702 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161711 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161730 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161750 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161761 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161882 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161898 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161907 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161911 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161949 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.161971 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.163651 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.163682 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:48 crc kubenswrapper[4954]: I1209 16:56:48.163699 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.167887 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85"} Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.167946 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053"} Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.167961 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8"} Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.167971 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f"} Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.167979 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a"} Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.167983 4954 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.168004 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.168025 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.167914 4954 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.168139 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.169217 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.169270 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.169282 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.169295 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.169299 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.169308 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.169272 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.169379 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.169394 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:49 crc kubenswrapper[4954]: I1209 16:56:49.218344 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:50 crc kubenswrapper[4954]: I1209 16:56:50.088236 4954 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 16:56:50 crc kubenswrapper[4954]: I1209 16:56:50.088370 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 16:56:50 crc kubenswrapper[4954]: I1209 16:56:50.170569 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:50 crc kubenswrapper[4954]: I1209 16:56:50.171057 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:50 crc kubenswrapper[4954]: I1209 16:56:50.171943 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:50 crc kubenswrapper[4954]: I1209 16:56:50.171990 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:50 crc kubenswrapper[4954]: I1209 16:56:50.172006 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:50 crc kubenswrapper[4954]: I1209 16:56:50.172057 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:50 crc kubenswrapper[4954]: I1209 16:56:50.172083 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:50 crc kubenswrapper[4954]: I1209 16:56:50.172098 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:52 crc kubenswrapper[4954]: I1209 16:56:52.013929 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:52 crc kubenswrapper[4954]: I1209 16:56:52.014213 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:52 crc kubenswrapper[4954]: I1209 16:56:52.016305 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:52 crc kubenswrapper[4954]: I1209 16:56:52.016368 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:52 crc kubenswrapper[4954]: I1209 16:56:52.016383 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:53 crc kubenswrapper[4954]: I1209 16:56:53.131394 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:56:53 crc kubenswrapper[4954]: I1209 16:56:53.131720 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:53 crc kubenswrapper[4954]: I1209 16:56:53.133343 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:53 crc kubenswrapper[4954]: I1209 16:56:53.133413 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:53 crc kubenswrapper[4954]: I1209 16:56:53.133430 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:53 crc kubenswrapper[4954]: I1209 16:56:53.276511 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 09 16:56:53 crc kubenswrapper[4954]: I1209 16:56:53.276840 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:53 crc kubenswrapper[4954]: I1209 16:56:53.278559 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:53 crc kubenswrapper[4954]: I1209 16:56:53.278649 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:53 crc kubenswrapper[4954]: I1209 16:56:53.278671 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:54 crc kubenswrapper[4954]: E1209 16:56:54.237874 4954 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 09 16:56:54 crc kubenswrapper[4954]: I1209 16:56:54.377084 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 16:56:54 crc kubenswrapper[4954]: I1209 16:56:54.377298 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:54 crc kubenswrapper[4954]: I1209 16:56:54.378926 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:54 crc kubenswrapper[4954]: I1209 16:56:54.378968 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:54 crc kubenswrapper[4954]: I1209 16:56:54.378985 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:55 crc kubenswrapper[4954]: I1209 16:56:55.472735 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 09 16:56:55 crc kubenswrapper[4954]: I1209 16:56:55.472880 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:55 crc kubenswrapper[4954]: I1209 16:56:55.474054 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:55 crc kubenswrapper[4954]: I1209 16:56:55.474081 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:55 crc kubenswrapper[4954]: I1209 16:56:55.474089 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:56:57 crc kubenswrapper[4954]: I1209 16:56:57.285223 4954 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 16:56:57 crc kubenswrapper[4954]: I1209 16:56:57.285290 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 16:56:57 crc kubenswrapper[4954]: I1209 16:56:57.290386 4954 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 09 16:56:57 crc kubenswrapper[4954]: I1209 16:56:57.290492 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 09 16:56:58 crc kubenswrapper[4954]: I1209 16:56:58.374998 4954 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 09 16:56:58 crc kubenswrapper[4954]: I1209 16:56:58.375084 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 09 16:56:59 crc kubenswrapper[4954]: I1209 16:56:59.223043 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:56:59 crc kubenswrapper[4954]: I1209 16:56:59.223208 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:56:59 crc kubenswrapper[4954]: I1209 16:56:59.224374 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:56:59 crc kubenswrapper[4954]: I1209 16:56:59.224420 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:56:59 crc kubenswrapper[4954]: I1209 16:56:59.224431 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:00 crc kubenswrapper[4954]: I1209 16:57:00.089199 4954 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 16:57:00 crc kubenswrapper[4954]: I1209 16:57:00.089311 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 16:57:02 crc kubenswrapper[4954]: E1209 16:57:02.289830 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.291870 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.292047 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.292519 4954 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.292571 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.293308 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.293363 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.293382 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.293686 4954 trace.go:236] Trace[1154280639]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 16:56:48.233) (total time: 14060ms): Dec 09 16:57:02 crc kubenswrapper[4954]: Trace[1154280639]: ---"Objects listed" error: 14060ms (16:57:02.293) Dec 09 16:57:02 crc kubenswrapper[4954]: Trace[1154280639]: [14.060187937s] [14.060187937s] END Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.293705 4954 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.293854 4954 trace.go:236] Trace[1237181923]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 16:56:47.426) (total time: 14867ms): Dec 09 16:57:02 crc kubenswrapper[4954]: Trace[1237181923]: ---"Objects listed" error: 14867ms (16:57:02.293) Dec 09 16:57:02 crc kubenswrapper[4954]: Trace[1237181923]: [14.867618122s] [14.867618122s] END Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.293885 4954 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 09 16:57:02 crc kubenswrapper[4954]: E1209 16:57:02.293989 4954 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.294926 4954 trace.go:236] Trace[451749356]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 16:56:47.488) (total time: 14806ms): Dec 09 16:57:02 crc kubenswrapper[4954]: Trace[451749356]: ---"Objects listed" error: 14806ms (16:57:02.294) Dec 09 16:57:02 crc kubenswrapper[4954]: Trace[451749356]: [14.806651208s] [14.806651208s] END Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.294997 4954 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.295047 4954 trace.go:236] Trace[1055877002]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 16:56:47.559) (total time: 14735ms): Dec 09 16:57:02 crc kubenswrapper[4954]: Trace[1055877002]: ---"Objects listed" error: 14735ms (16:57:02.294) Dec 09 16:57:02 crc kubenswrapper[4954]: Trace[1055877002]: [14.735175106s] [14.735175106s] END Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.295064 4954 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.296683 4954 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 09 16:57:02 crc kubenswrapper[4954]: I1209 16:57:02.297101 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.061149 4954 apiserver.go:52] "Watching apiserver" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.068646 4954 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.069052 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-p9vh6","openshift-multus/multus-h9n2h","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-node-7b4bw","openshift-machine-config-operator/machine-config-daemon-clkc5","openshift-multus/multus-additional-cni-plugins-x2fvb","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.069373 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.069486 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.069613 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.069628 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.069743 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.069938 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.070172 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.070434 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.070550 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.070646 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-p9vh6" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.070719 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.071047 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.071070 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.071282 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.072480 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.072489 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.074207 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.074227 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.074367 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.074491 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.074539 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.074715 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.074807 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.074838 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.075018 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.075044 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.075206 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.075367 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.075567 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.075698 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.075954 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.078500 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.078548 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.078659 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.078784 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.078809 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.078995 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.079060 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.079106 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.080182 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.080486 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.081708 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.081938 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.082117 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.082732 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.097263 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.110097 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.119672 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.128482 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.131981 4954 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.132048 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.143411 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.155078 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.165829 4954 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.167067 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.179544 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199526 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199580 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199637 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199665 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199690 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199720 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199793 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199822 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199846 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199865 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.199981 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200008 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200026 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200074 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200095 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200116 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200122 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200164 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200167 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200162 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200196 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.200256 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:57:03.700235018 +0000 UTC m=+20.088408838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200284 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200315 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200336 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200354 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200369 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200385 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200401 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200409 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200419 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200445 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200467 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200496 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200525 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200551 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200575 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200620 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200628 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200644 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200668 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200671 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200690 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200759 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200780 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200788 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200813 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200833 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200888 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200905 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200952 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200974 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201017 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201036 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201055 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201199 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201232 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201252 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201289 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201306 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201326 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201429 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201450 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201467 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201521 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201538 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201558 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201608 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201640 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200779 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.200925 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201020 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201133 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201153 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202390 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201188 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201199 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202419 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201236 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201333 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201353 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201386 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201424 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201490 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201522 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201616 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201647 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201706 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201801 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201842 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.201940 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202011 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202012 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202045 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202080 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202160 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202195 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202218 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202245 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202391 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202586 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.202713 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203021 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203446 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203492 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203516 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203542 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203566 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203609 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203631 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203655 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203680 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203702 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203726 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203749 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203770 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203790 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203814 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203838 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203861 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203888 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203917 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203946 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.203974 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204022 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204048 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204079 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204109 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204136 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204163 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204191 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204221 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204245 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204270 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204297 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204319 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204343 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204367 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204392 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204412 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204437 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204459 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204480 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204502 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204524 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204550 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204613 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204640 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204665 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204693 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204720 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204744 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204768 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204794 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204820 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204855 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204878 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204906 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204932 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204958 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.204981 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205003 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205025 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205047 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205068 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205090 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205097 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205115 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205176 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205195 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205215 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205231 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205259 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205285 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205323 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205358 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205402 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205436 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205462 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205488 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205524 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205550 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205574 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205619 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205647 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205670 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205699 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205725 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205750 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205777 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205802 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205826 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205852 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205880 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205906 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205930 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205961 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205987 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206020 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206043 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206068 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206092 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206118 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206141 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206163 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206187 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206213 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206239 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206262 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206286 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206312 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206340 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206410 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206438 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206467 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206492 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206521 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206547 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206571 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206615 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206651 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206678 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206717 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206742 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206768 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206793 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206822 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206848 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207153 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207220 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207250 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207277 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207303 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207326 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207350 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207375 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207398 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207511 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207611 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207642 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-var-lib-cni-bin\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207667 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5a6f4be9-a12a-4562-96b8-fd85694aa29c-mcd-auth-proxy-config\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207688 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-var-lib-kubelet\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207709 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-systemd-units\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207730 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-var-lib-openvswitch\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207756 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207778 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ea104037-eb12-4e58-8313-b35f63b7aa3c-hosts-file\") pod \"node-resolver-p9vh6\" (UID: \"ea104037-eb12-4e58-8313-b35f63b7aa3c\") " pod="openshift-dns/node-resolver-p9vh6" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207816 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-slash\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207839 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-env-overrides\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207871 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207894 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5a6f4be9-a12a-4562-96b8-fd85694aa29c-proxy-tls\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207918 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207963 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207991 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-node-log\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208019 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-conf-dir\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208042 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-etc-kubernetes\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208067 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-bin\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208089 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-config\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208118 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208141 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-kubelet\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208164 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msqm2\" (UniqueName: \"kubernetes.io/projected/5736742d-6a6c-4309-b95d-6103408901ec-kube-api-access-msqm2\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208192 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208242 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-cnibin\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208269 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-netns\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208313 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-log-socket\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208363 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-netd\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208502 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-system-cni-dir\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208528 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-run-multus-certs\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208806 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208839 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208892 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-os-release\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208919 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-ovn-kubernetes\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208968 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-script-lib\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.209002 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210316 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1eccf2e0-30a3-4201-a23f-85a92a522d72-cni-binary-copy\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210400 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5a6f4be9-a12a-4562-96b8-fd85694aa29c-rootfs\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210434 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210469 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-cni-binary-copy\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210500 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210529 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwdpm\" (UniqueName: \"kubernetes.io/projected/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-kube-api-access-nwdpm\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210555 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-cni-dir\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210581 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-socket-dir-parent\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210622 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-hostroot\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210654 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210681 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210706 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-cnibin\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210741 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-etc-openvswitch\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210766 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-openvswitch\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210791 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5f67\" (UniqueName: \"kubernetes.io/projected/5a6f4be9-a12a-4562-96b8-fd85694aa29c-kube-api-access-f5f67\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210814 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-run-netns\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210834 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-ovn\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210863 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz6xg\" (UniqueName: \"kubernetes.io/projected/ea104037-eb12-4e58-8313-b35f63b7aa3c-kube-api-access-vz6xg\") pod \"node-resolver-p9vh6\" (UID: \"ea104037-eb12-4e58-8313-b35f63b7aa3c\") " pod="openshift-dns/node-resolver-p9vh6" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210885 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-system-cni-dir\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210907 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-run-k8s-cni-cncf-io\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210946 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cs9t\" (UniqueName: \"kubernetes.io/projected/1eccf2e0-30a3-4201-a23f-85a92a522d72-kube-api-access-7cs9t\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.210968 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-systemd\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211000 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211032 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211056 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211078 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-var-lib-cni-multus\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211100 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-daemon-config\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211123 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5736742d-6a6c-4309-b95d-6103408901ec-ovn-node-metrics-cert\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211148 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-os-release\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211252 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211273 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211292 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211320 4954 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211336 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211351 4954 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211367 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211383 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211400 4954 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211414 4954 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211429 4954 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211443 4954 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211459 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211474 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211489 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211504 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211532 4954 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211547 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211562 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211576 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211638 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211655 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211669 4954 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211683 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211703 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211718 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211735 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211749 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211764 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211779 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211796 4954 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211813 4954 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211828 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211890 4954 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211927 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211945 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211964 4954 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211979 4954 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211994 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.212008 4954 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.212023 4954 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.212039 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.215737 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205402 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205468 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205759 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206192 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206272 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206427 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206761 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.206813 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207023 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207161 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207182 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207200 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.207335 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.208880 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.209126 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.209298 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.209941 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.209951 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211490 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.220901 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211757 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.211784 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.212223 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.212441 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.212475 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.212815 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.212821 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.212993 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.213289 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.213650 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.214103 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.214386 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.214420 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.214677 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.214766 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.215036 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.215266 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.221039 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.215414 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.215584 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.216211 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.216227 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.216278 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.221272 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.221309 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.216403 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.216488 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.216720 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.216731 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.217084 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.217307 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.217421 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.217476 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.209376 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.217633 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.217828 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.217994 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.218014 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.218010 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.219454 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.219673 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.220211 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.220732 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.221718 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.222225 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.222553 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.222818 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.223036 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.223920 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.224019 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.215618 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.224670 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.225193 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.225483 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.225789 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.226021 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.226396 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.226640 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.226665 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.226707 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.226888 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.227161 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.227263 4954 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.227294 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b"} Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.227642 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.227809 4954 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.227880 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:03.727855674 +0000 UTC m=+20.116029494 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.227973 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.228318 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.228892 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.229263 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.229787 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.230154 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.230387 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.230644 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.230980 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.233414 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.233763 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.234096 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.235843 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.237917 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.238033 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.238251 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.239570 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.228036 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.240699 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.240727 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.227267 4954 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b" exitCode=255 Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.241052 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.241575 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.241678 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.241727 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.242112 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.242140 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.242211 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.242368 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.242702 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.242658 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.243176 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.243321 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.244120 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.244329 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.244348 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.244466 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.244508 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.244527 4954 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.244483 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.244648 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:03.74462313 +0000 UTC m=+20.132797170 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.244722 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.244792 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.244979 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.245058 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.245137 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.205946 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.245435 4954 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.245508 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:03.745491478 +0000 UTC m=+20.133665298 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.245516 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.245735 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.246223 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.246270 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.246449 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.246744 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.246907 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.247333 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.247654 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.248016 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.248053 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.248216 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.248406 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.248540 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.251146 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.251187 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.251202 4954 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.251252 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:03.751231987 +0000 UTC m=+20.139405807 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.251896 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.252079 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.255360 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.256554 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.256686 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.257151 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.259871 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.260287 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.261366 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.261485 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.261538 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.273871 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.274418 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.274772 4954 scope.go:117] "RemoveContainer" containerID="9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.274227 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.275237 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.275266 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.275278 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.276315 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.275822 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.275893 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.289628 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.295161 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.312836 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313214 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-system-cni-dir\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313239 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-run-multus-certs\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313255 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-netns\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313271 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-log-socket\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313285 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-netd\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313298 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313312 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-os-release\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313326 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-ovn-kubernetes\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313340 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-script-lib\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313356 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-cni-binary-copy\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313374 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1eccf2e0-30a3-4201-a23f-85a92a522d72-cni-binary-copy\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313388 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5a6f4be9-a12a-4562-96b8-fd85694aa29c-rootfs\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313402 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-hostroot\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313422 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-cnibin\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313436 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwdpm\" (UniqueName: \"kubernetes.io/projected/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-kube-api-access-nwdpm\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313451 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-cni-dir\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313466 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-socket-dir-parent\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313481 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-etc-openvswitch\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313496 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-openvswitch\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313510 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5f67\" (UniqueName: \"kubernetes.io/projected/5a6f4be9-a12a-4562-96b8-fd85694aa29c-kube-api-access-f5f67\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313525 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz6xg\" (UniqueName: \"kubernetes.io/projected/ea104037-eb12-4e58-8313-b35f63b7aa3c-kube-api-access-vz6xg\") pod \"node-resolver-p9vh6\" (UID: \"ea104037-eb12-4e58-8313-b35f63b7aa3c\") " pod="openshift-dns/node-resolver-p9vh6" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313538 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-system-cni-dir\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313551 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-run-netns\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313564 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-ovn\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313583 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313626 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-run-k8s-cni-cncf-io\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313642 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cs9t\" (UniqueName: \"kubernetes.io/projected/1eccf2e0-30a3-4201-a23f-85a92a522d72-kube-api-access-7cs9t\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313659 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-systemd\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313692 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-os-release\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313712 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-var-lib-cni-multus\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313730 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-daemon-config\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313753 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5736742d-6a6c-4309-b95d-6103408901ec-ovn-node-metrics-cert\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313776 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313797 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-var-lib-cni-bin\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313823 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5a6f4be9-a12a-4562-96b8-fd85694aa29c-mcd-auth-proxy-config\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313843 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-var-lib-kubelet\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313861 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-systemd-units\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313881 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-var-lib-openvswitch\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313904 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313929 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ea104037-eb12-4e58-8313-b35f63b7aa3c-hosts-file\") pod \"node-resolver-p9vh6\" (UID: \"ea104037-eb12-4e58-8313-b35f63b7aa3c\") " pod="openshift-dns/node-resolver-p9vh6" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313971 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-slash\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.313993 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-env-overrides\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314016 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5a6f4be9-a12a-4562-96b8-fd85694aa29c-proxy-tls\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314048 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-node-log\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314036 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-openvswitch\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314034 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314118 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-conf-dir\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314237 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-log-socket\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314071 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-conf-dir\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314295 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-run-multus-certs\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314325 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-etc-kubernetes\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314296 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-etc-kubernetes\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314354 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-system-cni-dir\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314363 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-bin\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314390 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-config\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314390 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-cnibin\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314415 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314443 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-cnibin\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314479 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-kubelet\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314511 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msqm2\" (UniqueName: \"kubernetes.io/projected/5736742d-6a6c-4309-b95d-6103408901ec-kube-api-access-msqm2\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314583 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314618 4954 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314632 4954 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314643 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314656 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314667 4954 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314678 4954 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314690 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314702 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314712 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314722 4954 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314793 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314807 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314819 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314829 4954 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314842 4954 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314852 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314863 4954 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314874 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314885 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314896 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314908 4954 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314919 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314932 4954 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314944 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314956 4954 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314957 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-script-lib\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315007 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-netns\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315045 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-bin\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315147 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1eccf2e0-30a3-4201-a23f-85a92a522d72-cni-binary-copy\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315152 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-netd\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315170 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-cni-binary-copy\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315197 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5a6f4be9-a12a-4562-96b8-fd85694aa29c-rootfs\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315230 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-hostroot\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315567 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315677 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-cni-dir\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315749 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-socket-dir-parent\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315789 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-etc-openvswitch\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315805 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-config\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315829 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315850 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-os-release\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315863 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-var-lib-cni-bin\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315828 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315887 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-ovn-kubernetes\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315910 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-var-lib-kubelet\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.314967 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315937 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-system-cni-dir\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315946 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315965 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315972 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315981 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315987 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-ovn\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.315966 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-run-netns\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316005 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316023 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316030 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316038 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316053 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316057 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-run-k8s-cni-cncf-io\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316070 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316087 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316102 4954 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316115 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316129 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316143 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316156 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316171 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-systemd\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316174 4954 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316214 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316214 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-cnibin\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316225 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316239 4954 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316254 4954 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316021 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-systemd-units\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316267 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316281 4954 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316293 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316313 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-slash\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316348 4954 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316385 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ea104037-eb12-4e58-8313-b35f63b7aa3c-hosts-file\") pod \"node-resolver-p9vh6\" (UID: \"ea104037-eb12-4e58-8313-b35f63b7aa3c\") " pod="openshift-dns/node-resolver-p9vh6" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316415 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5a6f4be9-a12a-4562-96b8-fd85694aa29c-mcd-auth-proxy-config\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316434 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-os-release\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316469 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1eccf2e0-30a3-4201-a23f-85a92a522d72-host-var-lib-cni-multus\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316490 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316554 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-env-overrides\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316051 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-var-lib-openvswitch\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316254 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-kubelet\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316673 4954 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316700 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-node-log\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316718 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316733 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316747 4954 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316763 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316776 4954 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316852 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316867 4954 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316879 4954 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316902 4954 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316916 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316929 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316944 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316958 4954 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316971 4954 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316985 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.316997 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317011 4954 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317024 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317035 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317047 4954 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317059 4954 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317071 4954 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317082 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317093 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317104 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317118 4954 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317131 4954 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317144 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317158 4954 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317169 4954 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317181 4954 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317196 4954 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317181 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1eccf2e0-30a3-4201-a23f-85a92a522d72-multus-daemon-config\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317270 4954 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317291 4954 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317304 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317320 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317333 4954 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317346 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317359 4954 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317371 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317383 4954 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317396 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317409 4954 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317421 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317437 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317451 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317464 4954 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317476 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317489 4954 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317502 4954 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317515 4954 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317530 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317542 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317554 4954 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317567 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317581 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317617 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317631 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317642 4954 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317653 4954 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317664 4954 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317675 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317687 4954 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317706 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317717 4954 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317726 4954 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317736 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317747 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317758 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317768 4954 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317778 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317790 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317801 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317812 4954 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317822 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317833 4954 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317843 4954 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317853 4954 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317862 4954 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317871 4954 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317882 4954 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317892 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317903 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317914 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317924 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317934 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317943 4954 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317954 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317967 4954 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317978 4954 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317988 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.317999 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.318010 4954 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.318022 4954 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.318034 4954 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.318051 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.318064 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.318076 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.328203 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5a6f4be9-a12a-4562-96b8-fd85694aa29c-proxy-tls\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.332853 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.333292 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5736742d-6a6c-4309-b95d-6103408901ec-ovn-node-metrics-cert\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.337092 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.339392 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msqm2\" (UniqueName: \"kubernetes.io/projected/5736742d-6a6c-4309-b95d-6103408901ec-kube-api-access-msqm2\") pod \"ovnkube-node-7b4bw\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.339806 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5f67\" (UniqueName: \"kubernetes.io/projected/5a6f4be9-a12a-4562-96b8-fd85694aa29c-kube-api-access-f5f67\") pod \"machine-config-daemon-clkc5\" (UID: \"5a6f4be9-a12a-4562-96b8-fd85694aa29c\") " pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.344510 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwdpm\" (UniqueName: \"kubernetes.io/projected/eb6c6493-2ebd-4f44-b988-2722f2f44eb5-kube-api-access-nwdpm\") pod \"multus-additional-cni-plugins-x2fvb\" (UID: \"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\") " pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.346714 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cs9t\" (UniqueName: \"kubernetes.io/projected/1eccf2e0-30a3-4201-a23f-85a92a522d72-kube-api-access-7cs9t\") pod \"multus-h9n2h\" (UID: \"1eccf2e0-30a3-4201-a23f-85a92a522d72\") " pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.348505 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz6xg\" (UniqueName: \"kubernetes.io/projected/ea104037-eb12-4e58-8313-b35f63b7aa3c-kube-api-access-vz6xg\") pod \"node-resolver-p9vh6\" (UID: \"ea104037-eb12-4e58-8313-b35f63b7aa3c\") " pod="openshift-dns/node-resolver-p9vh6" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.352246 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.365391 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.374321 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.385420 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.388105 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.394938 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.395883 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.402650 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.407712 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.409397 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-p9vh6" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.419547 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h9n2h" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.420054 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.420086 4954 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 09 16:57:03 crc kubenswrapper[4954]: W1209 16:57:03.425147 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-835bbd75c942be1cb9a7ba7cee813f9b6de77ebbeed7b149421e7400ebc3810a WatchSource:0}: Error finding container 835bbd75c942be1cb9a7ba7cee813f9b6de77ebbeed7b149421e7400ebc3810a: Status 404 returned error can't find the container with id 835bbd75c942be1cb9a7ba7cee813f9b6de77ebbeed7b149421e7400ebc3810a Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.425092 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.427006 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 16:57:03 crc kubenswrapper[4954]: W1209 16:57:03.432194 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-e50025d59a7c276a402f66c6b13db29b0e98c852969f2251d021d8085ec4657e WatchSource:0}: Error finding container e50025d59a7c276a402f66c6b13db29b0e98c852969f2251d021d8085ec4657e: Status 404 returned error can't find the container with id e50025d59a7c276a402f66c6b13db29b0e98c852969f2251d021d8085ec4657e Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.434492 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.443003 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.443021 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.452644 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: W1209 16:57:03.460583 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb6c6493_2ebd_4f44_b988_2722f2f44eb5.slice/crio-ed488a1d383a4bebaad330de9244183b04ada2af861d0a93d284f18f39627dbf WatchSource:0}: Error finding container ed488a1d383a4bebaad330de9244183b04ada2af861d0a93d284f18f39627dbf: Status 404 returned error can't find the container with id ed488a1d383a4bebaad330de9244183b04ada2af861d0a93d284f18f39627dbf Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.467682 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.479713 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: W1209 16:57:03.486175 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a6f4be9_a12a_4562_96b8_fd85694aa29c.slice/crio-5fcd44b9fa04efad03f6b81ac67485de0f4809e01d5c89c7b760ba91c818ee0b WatchSource:0}: Error finding container 5fcd44b9fa04efad03f6b81ac67485de0f4809e01d5c89c7b760ba91c818ee0b: Status 404 returned error can't find the container with id 5fcd44b9fa04efad03f6b81ac67485de0f4809e01d5c89c7b760ba91c818ee0b Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.490027 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.501636 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:03 crc kubenswrapper[4954]: W1209 16:57:03.518195 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5736742d_6a6c_4309_b95d_6103408901ec.slice/crio-62df6ac04ca23854487e97a4cb353c94a24264ca3a1ab3705570624c599af8f6 WatchSource:0}: Error finding container 62df6ac04ca23854487e97a4cb353c94a24264ca3a1ab3705570624c599af8f6: Status 404 returned error can't find the container with id 62df6ac04ca23854487e97a4cb353c94a24264ca3a1ab3705570624c599af8f6 Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.721946 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.722100 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:57:04.722075742 +0000 UTC m=+21.110249562 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.823243 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.823294 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.823317 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:03 crc kubenswrapper[4954]: I1209 16:57:03.823345 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.823462 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.823476 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.823487 4954 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.823536 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:04.823521603 +0000 UTC m=+21.211695423 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.823869 4954 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.823884 4954 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.823920 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:04.82390989 +0000 UTC m=+21.212083710 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.823975 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:04.823952952 +0000 UTC m=+21.212126772 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.823903 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.823996 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.824005 4954 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:03 crc kubenswrapper[4954]: E1209 16:57:03.824038 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:04.824026425 +0000 UTC m=+21.212200285 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.123662 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.124818 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.125982 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.126704 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.127763 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.128362 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.129005 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.130716 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.131420 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.133061 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.133421 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.133784 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.134993 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.135567 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.136184 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.141242 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.142007 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.142976 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.143391 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.143466 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.144125 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.144786 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.145301 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.145888 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.146325 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.152056 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.152630 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.153708 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.154365 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.155412 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.156191 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.157045 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.157099 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.157553 4954 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.157672 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.159460 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.160379 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.160826 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.162737 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.164017 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.164725 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.166010 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.166851 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.167960 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.168692 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.169674 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.169879 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.170638 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.171568 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.172205 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.173357 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.174344 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.175379 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.176149 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.177325 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.178029 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.178691 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.179689 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.199972 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.217891 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.229530 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.245893 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981" exitCode=0 Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.245989 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.246026 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"62df6ac04ca23854487e97a4cb353c94a24264ca3a1ab3705570624c599af8f6"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.247405 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"835bbd75c942be1cb9a7ba7cee813f9b6de77ebbeed7b149421e7400ebc3810a"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.248576 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.249942 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.250004 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.250019 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"5fcd44b9fa04efad03f6b81ac67485de0f4809e01d5c89c7b760ba91c818ee0b"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.257248 4954 generic.go:334] "Generic (PLEG): container finished" podID="eb6c6493-2ebd-4f44-b988-2722f2f44eb5" containerID="39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d" exitCode=0 Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.257395 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" event={"ID":"eb6c6493-2ebd-4f44-b988-2722f2f44eb5","Type":"ContainerDied","Data":"39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.257433 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" event={"ID":"eb6c6493-2ebd-4f44-b988-2722f2f44eb5","Type":"ContainerStarted","Data":"ed488a1d383a4bebaad330de9244183b04ada2af861d0a93d284f18f39627dbf"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.262701 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.262759 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.262770 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e50025d59a7c276a402f66c6b13db29b0e98c852969f2251d021d8085ec4657e"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.264060 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.264131 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6563adae19f5740332b1be1a663f705f5762ffd6c1712df275e46d6170f5c1a9"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.266936 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.267548 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9n2h" event={"ID":"1eccf2e0-30a3-4201-a23f-85a92a522d72","Type":"ContainerStarted","Data":"3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.267623 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9n2h" event={"ID":"1eccf2e0-30a3-4201-a23f-85a92a522d72","Type":"ContainerStarted","Data":"9b60a9459c90dea62a5521625e4ebd1f7ab1f64e25efe7c97e90960f9a1b20bc"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.270631 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-p9vh6" event={"ID":"ea104037-eb12-4e58-8313-b35f63b7aa3c","Type":"ContainerStarted","Data":"78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.270709 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-p9vh6" event={"ID":"ea104037-eb12-4e58-8313-b35f63b7aa3c","Type":"ContainerStarted","Data":"28a4235345bf6fa4c9a1f2966ad040df641b03a0c233139867ed066da9deb0b8"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.273348 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.275064 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97"} Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.275555 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.280639 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.295161 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.316560 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.323035 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-r4gn7"] Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.323441 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.326897 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.326914 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.327238 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.328952 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.335449 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.349675 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.363536 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.381193 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.396159 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.411002 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.425979 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.430779 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53c57bd8-d348-441c-b1b9-3e1e493d024c-host\") pod \"node-ca-r4gn7\" (UID: \"53c57bd8-d348-441c-b1b9-3e1e493d024c\") " pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.430845 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrlqb\" (UniqueName: \"kubernetes.io/projected/53c57bd8-d348-441c-b1b9-3e1e493d024c-kube-api-access-mrlqb\") pod \"node-ca-r4gn7\" (UID: \"53c57bd8-d348-441c-b1b9-3e1e493d024c\") " pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.430876 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/53c57bd8-d348-441c-b1b9-3e1e493d024c-serviceca\") pod \"node-ca-r4gn7\" (UID: \"53c57bd8-d348-441c-b1b9-3e1e493d024c\") " pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.440260 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.454484 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.470340 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.484967 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.500122 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.531265 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.531498 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53c57bd8-d348-441c-b1b9-3e1e493d024c-host\") pod \"node-ca-r4gn7\" (UID: \"53c57bd8-d348-441c-b1b9-3e1e493d024c\") " pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.531607 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrlqb\" (UniqueName: \"kubernetes.io/projected/53c57bd8-d348-441c-b1b9-3e1e493d024c-kube-api-access-mrlqb\") pod \"node-ca-r4gn7\" (UID: \"53c57bd8-d348-441c-b1b9-3e1e493d024c\") " pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.531653 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/53c57bd8-d348-441c-b1b9-3e1e493d024c-serviceca\") pod \"node-ca-r4gn7\" (UID: \"53c57bd8-d348-441c-b1b9-3e1e493d024c\") " pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.531673 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53c57bd8-d348-441c-b1b9-3e1e493d024c-host\") pod \"node-ca-r4gn7\" (UID: \"53c57bd8-d348-441c-b1b9-3e1e493d024c\") " pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.532825 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/53c57bd8-d348-441c-b1b9-3e1e493d024c-serviceca\") pod \"node-ca-r4gn7\" (UID: \"53c57bd8-d348-441c-b1b9-3e1e493d024c\") " pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.548993 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.556343 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrlqb\" (UniqueName: \"kubernetes.io/projected/53c57bd8-d348-441c-b1b9-3e1e493d024c-kube-api-access-mrlqb\") pod \"node-ca-r4gn7\" (UID: \"53c57bd8-d348-441c-b1b9-3e1e493d024c\") " pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.560914 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-r4gn7" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.566562 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: W1209 16:57:04.583789 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53c57bd8_d348_441c_b1b9_3e1e493d024c.slice/crio-686577e5b5a094bcc9561a88e0ea3609cdda6ddb19f95c4d98a18c44836af37a WatchSource:0}: Error finding container 686577e5b5a094bcc9561a88e0ea3609cdda6ddb19f95c4d98a18c44836af37a: Status 404 returned error can't find the container with id 686577e5b5a094bcc9561a88e0ea3609cdda6ddb19f95c4d98a18c44836af37a Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.586370 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.602496 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.619034 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.638354 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.656927 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.672985 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.696260 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.717358 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.732520 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.732611 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.732751 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:57:06.732735288 +0000 UTC m=+23.120909098 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.747387 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.836166 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.836243 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.836289 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:04 crc kubenswrapper[4954]: I1209 16:57:04.836322 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836508 4954 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836619 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:06.836572563 +0000 UTC m=+23.224746383 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836624 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836669 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836681 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836708 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836725 4954 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836728 4954 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836688 4954 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836762 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:06.836752541 +0000 UTC m=+23.224926371 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.836818 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:06.836801733 +0000 UTC m=+23.224975553 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:04 crc kubenswrapper[4954]: E1209 16:57:04.837664 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:06.837589877 +0000 UTC m=+23.225763837 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.120202 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.120222 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.120289 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:05 crc kubenswrapper[4954]: E1209 16:57:05.120988 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:05 crc kubenswrapper[4954]: E1209 16:57:05.121085 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:05 crc kubenswrapper[4954]: E1209 16:57:05.121133 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.280966 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-r4gn7" event={"ID":"53c57bd8-d348-441c-b1b9-3e1e493d024c","Type":"ContainerStarted","Data":"b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.281031 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-r4gn7" event={"ID":"53c57bd8-d348-441c-b1b9-3e1e493d024c","Type":"ContainerStarted","Data":"686577e5b5a094bcc9561a88e0ea3609cdda6ddb19f95c4d98a18c44836af37a"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.285111 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.285153 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.285168 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.285181 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.287258 4954 generic.go:334] "Generic (PLEG): container finished" podID="eb6c6493-2ebd-4f44-b988-2722f2f44eb5" containerID="5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91" exitCode=0 Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.287339 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" event={"ID":"eb6c6493-2ebd-4f44-b988-2722f2f44eb5","Type":"ContainerDied","Data":"5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.301259 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.312493 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.324080 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.351666 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.367766 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.384313 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.399675 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.412932 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.428227 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.444983 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.463088 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.476582 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.494069 4954 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.494196 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.495820 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.495865 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.495878 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.495996 4954 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.500087 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.505730 4954 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.506099 4954 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.507663 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.507710 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.507721 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.507745 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.507758 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:05Z","lastTransitionTime":"2025-12-09T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.510325 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.513450 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.516093 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.527314 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: E1209 16:57:05.529446 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.537236 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.537294 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.537310 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.537339 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.537353 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:05Z","lastTransitionTime":"2025-12-09T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.551917 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: E1209 16:57:05.552688 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.556625 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.556665 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.556675 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.556690 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.556703 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:05Z","lastTransitionTime":"2025-12-09T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:05 crc kubenswrapper[4954]: E1209 16:57:05.571060 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.571416 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.575940 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.575974 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.575987 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.576005 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.576017 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:05Z","lastTransitionTime":"2025-12-09T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:05 crc kubenswrapper[4954]: E1209 16:57:05.590539 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.593670 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.594546 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.594632 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.594648 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.594671 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.594684 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:05Z","lastTransitionTime":"2025-12-09T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:05 crc kubenswrapper[4954]: E1209 16:57:05.605889 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: E1209 16:57:05.606015 4954 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.607676 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.607958 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.608002 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.608016 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.608035 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.608051 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:05Z","lastTransitionTime":"2025-12-09T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.622195 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.638565 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.650348 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.671010 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.685323 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.696632 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.707533 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.709982 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.710011 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.710021 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.710033 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.710043 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:05Z","lastTransitionTime":"2025-12-09T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.720817 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.730982 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.740758 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.773359 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.811923 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.812689 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.812735 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.812747 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.812776 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.812785 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:05Z","lastTransitionTime":"2025-12-09T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.852332 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.892881 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.916792 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.917002 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.917016 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.917045 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.917056 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:05Z","lastTransitionTime":"2025-12-09T16:57:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.931085 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:05 crc kubenswrapper[4954]: I1209 16:57:05.969730 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:05Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.020202 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.020243 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.020252 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.020270 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.020282 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:06Z","lastTransitionTime":"2025-12-09T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.027351 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.057614 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.092640 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.123057 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.123114 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.123124 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.123145 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.123160 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:06Z","lastTransitionTime":"2025-12-09T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.133881 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.179003 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.225780 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.225824 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.225834 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.225852 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.225864 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:06Z","lastTransitionTime":"2025-12-09T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.293948 4954 generic.go:334] "Generic (PLEG): container finished" podID="eb6c6493-2ebd-4f44-b988-2722f2f44eb5" containerID="c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0" exitCode=0 Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.294028 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" event={"ID":"eb6c6493-2ebd-4f44-b988-2722f2f44eb5","Type":"ContainerDied","Data":"c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.298518 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.298668 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.320329 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.328918 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.328944 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.328956 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.328970 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.328980 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:06Z","lastTransitionTime":"2025-12-09T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.333863 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.347669 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.362441 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.375933 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.413889 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.431968 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.432018 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.432027 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.432049 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.432065 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:06Z","lastTransitionTime":"2025-12-09T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.457670 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.495625 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.530417 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.534607 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.534636 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.534646 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.534659 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.534669 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:06Z","lastTransitionTime":"2025-12-09T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.583657 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.614555 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.637303 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.637342 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.637351 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.637364 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.637374 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:06Z","lastTransitionTime":"2025-12-09T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.651303 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.697158 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.740608 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.740642 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.740657 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.740704 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.740717 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:06Z","lastTransitionTime":"2025-12-09T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.740809 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:06Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.755313 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.755561 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:57:10.755532184 +0000 UTC m=+27.143706014 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.843449 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.843486 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.843495 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.843511 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.843522 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:06Z","lastTransitionTime":"2025-12-09T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.856170 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.856211 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.856242 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.856264 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856327 4954 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856352 4954 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856407 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:10.856386771 +0000 UTC m=+27.244560591 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856404 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856421 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:10.856415422 +0000 UTC m=+27.244589242 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856427 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856437 4954 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856437 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856467 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856483 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:10.856470005 +0000 UTC m=+27.244643825 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856484 4954 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:06 crc kubenswrapper[4954]: E1209 16:57:06.856541 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:10.856523677 +0000 UTC m=+27.244697537 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.945975 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.946015 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.946024 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.946037 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:06 crc kubenswrapper[4954]: I1209 16:57:06.946079 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:06Z","lastTransitionTime":"2025-12-09T16:57:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.049421 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.049475 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.049491 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.049510 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.049521 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:07Z","lastTransitionTime":"2025-12-09T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.093000 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.096667 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.102019 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.106990 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.117027 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.119247 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.119305 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:07 crc kubenswrapper[4954]: E1209 16:57:07.119346 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.119248 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:07 crc kubenswrapper[4954]: E1209 16:57:07.119454 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:07 crc kubenswrapper[4954]: E1209 16:57:07.119542 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.139757 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.150888 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.151553 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.151585 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.151610 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.151627 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.151638 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:07Z","lastTransitionTime":"2025-12-09T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.160417 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.169323 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.186001 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.198413 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.208468 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.216840 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.227732 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.238967 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.253558 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.253615 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.253625 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.253641 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.253651 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:07Z","lastTransitionTime":"2025-12-09T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.271780 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.303122 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.305562 4954 generic.go:334] "Generic (PLEG): container finished" podID="eb6c6493-2ebd-4f44-b988-2722f2f44eb5" containerID="8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa" exitCode=0 Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.305624 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" event={"ID":"eb6c6493-2ebd-4f44-b988-2722f2f44eb5","Type":"ContainerDied","Data":"8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.318141 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.351583 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.361039 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.361080 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.361090 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.361106 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.361117 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:07Z","lastTransitionTime":"2025-12-09T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.390168 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.433858 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.463172 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.463203 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.463212 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.463226 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.463236 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:07Z","lastTransitionTime":"2025-12-09T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.472373 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.512257 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.552733 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.566819 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.566856 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.566866 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.566881 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.566891 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:07Z","lastTransitionTime":"2025-12-09T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.600229 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.639759 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.669893 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.669936 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.669951 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.669970 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.669981 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:07Z","lastTransitionTime":"2025-12-09T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.675831 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.714214 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.751046 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.772100 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.772162 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.772173 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.772190 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.772201 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:07Z","lastTransitionTime":"2025-12-09T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.793836 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.832717 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.874263 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.874307 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.874319 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.874338 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.874351 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:07Z","lastTransitionTime":"2025-12-09T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.875356 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.913309 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.976634 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.976679 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.976697 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.976751 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:07 crc kubenswrapper[4954]: I1209 16:57:07.976765 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:07Z","lastTransitionTime":"2025-12-09T16:57:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.078729 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.078769 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.078782 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.078798 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.078807 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:08Z","lastTransitionTime":"2025-12-09T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.181015 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.181053 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.181061 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.181077 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.181088 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:08Z","lastTransitionTime":"2025-12-09T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.283087 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.283119 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.283127 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.283139 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.283148 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:08Z","lastTransitionTime":"2025-12-09T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.312129 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.314930 4954 generic.go:334] "Generic (PLEG): container finished" podID="eb6c6493-2ebd-4f44-b988-2722f2f44eb5" containerID="92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40" exitCode=0 Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.315331 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" event={"ID":"eb6c6493-2ebd-4f44-b988-2722f2f44eb5","Type":"ContainerDied","Data":"92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.343421 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.357676 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.369500 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.382288 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.384965 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.385007 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.385017 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.385032 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.385042 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:08Z","lastTransitionTime":"2025-12-09T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.400972 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.415479 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.428234 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.439456 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.452106 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.462935 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.475371 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.487606 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.487640 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.487650 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.487664 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.487673 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:08Z","lastTransitionTime":"2025-12-09T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.490054 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.504096 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.518255 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.529221 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:08Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.590655 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.590717 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.590727 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.590749 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.590770 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:08Z","lastTransitionTime":"2025-12-09T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.693869 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.693971 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.693991 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.694016 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.694034 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:08Z","lastTransitionTime":"2025-12-09T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.797161 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.797221 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.797241 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.797267 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.797287 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:08Z","lastTransitionTime":"2025-12-09T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.899505 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.899545 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.899557 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.899572 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:08 crc kubenswrapper[4954]: I1209 16:57:08.899583 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:08Z","lastTransitionTime":"2025-12-09T16:57:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.002549 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.002624 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.002635 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.002656 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.002667 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:09Z","lastTransitionTime":"2025-12-09T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.105774 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.106012 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.106129 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.106246 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.106394 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:09Z","lastTransitionTime":"2025-12-09T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.119215 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.119271 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:09 crc kubenswrapper[4954]: E1209 16:57:09.119312 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.119217 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:09 crc kubenswrapper[4954]: E1209 16:57:09.119672 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:09 crc kubenswrapper[4954]: E1209 16:57:09.119965 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.209005 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.209052 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.209064 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.209080 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.209090 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:09Z","lastTransitionTime":"2025-12-09T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.312813 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.313103 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.313235 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.313337 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.313397 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:09Z","lastTransitionTime":"2025-12-09T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.321234 4954 generic.go:334] "Generic (PLEG): container finished" podID="eb6c6493-2ebd-4f44-b988-2722f2f44eb5" containerID="713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c" exitCode=0 Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.321287 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" event={"ID":"eb6c6493-2ebd-4f44-b988-2722f2f44eb5","Type":"ContainerDied","Data":"713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.336543 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.347991 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.359488 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.374728 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.390171 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.407993 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.415835 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.415873 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.415886 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.415913 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.415929 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:09Z","lastTransitionTime":"2025-12-09T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.425773 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.440489 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.456042 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.470710 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.484712 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.498341 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.513382 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.518067 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.518102 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.518111 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.518127 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.518138 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:09Z","lastTransitionTime":"2025-12-09T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.531610 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.549719 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:09Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.620614 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.620652 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.620664 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.620679 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.620689 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:09Z","lastTransitionTime":"2025-12-09T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.723262 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.723520 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.723528 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.723542 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.723550 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:09Z","lastTransitionTime":"2025-12-09T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.826138 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.826174 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.826184 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.826197 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.826205 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:09Z","lastTransitionTime":"2025-12-09T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.928099 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.928128 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.928136 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.928148 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:09 crc kubenswrapper[4954]: I1209 16:57:09.928158 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:09Z","lastTransitionTime":"2025-12-09T16:57:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.033917 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.033967 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.033980 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.034375 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.034417 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:10Z","lastTransitionTime":"2025-12-09T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.136979 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.137032 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.137043 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.137062 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.137073 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:10Z","lastTransitionTime":"2025-12-09T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.239744 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.239789 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.239799 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.239820 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.239834 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:10Z","lastTransitionTime":"2025-12-09T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.330295 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.331102 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.331177 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.337765 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" event={"ID":"eb6c6493-2ebd-4f44-b988-2722f2f44eb5","Type":"ContainerStarted","Data":"f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.342804 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.342866 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.342877 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.342937 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.342959 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:10Z","lastTransitionTime":"2025-12-09T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.353540 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.356759 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.359089 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.369666 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.382682 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.397310 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.416898 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.432310 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.445761 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.445803 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.445811 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.445828 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.445856 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:10Z","lastTransitionTime":"2025-12-09T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.448153 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.460730 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.477055 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.491042 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.511207 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.526130 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.548985 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.549033 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.549048 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.549074 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.549091 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:10Z","lastTransitionTime":"2025-12-09T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.549979 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.574556 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.590992 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.604938 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.621321 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.636041 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.652075 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.652166 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.652183 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.652204 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.652221 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:10Z","lastTransitionTime":"2025-12-09T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.654714 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.670191 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.684161 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.695062 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.713060 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.725763 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.737547 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.754177 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.754227 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.754239 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.754257 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.754275 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:10Z","lastTransitionTime":"2025-12-09T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.756821 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.776726 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.790127 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.795536 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.795642 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:57:18.795623129 +0000 UTC m=+35.183796949 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.804566 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.816225 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:10Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.857026 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.857068 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.857077 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.857100 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.857115 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:10Z","lastTransitionTime":"2025-12-09T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.897244 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.897301 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.897332 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.897356 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897450 4954 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897532 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897553 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897553 4954 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897564 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:18.897537471 +0000 UTC m=+35.285711291 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897675 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897723 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897740 4954 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897691 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:18.897670337 +0000 UTC m=+35.285844357 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897567 4954 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897882 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:18.897824703 +0000 UTC m=+35.285998523 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:10 crc kubenswrapper[4954]: E1209 16:57:10.897944 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:18.897927148 +0000 UTC m=+35.286100978 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.960582 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.960654 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.960668 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.960689 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:10 crc kubenswrapper[4954]: I1209 16:57:10.960708 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:10Z","lastTransitionTime":"2025-12-09T16:57:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.064631 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.064677 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.064691 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.064715 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.064730 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:11Z","lastTransitionTime":"2025-12-09T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.120137 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.120285 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.120335 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:11 crc kubenswrapper[4954]: E1209 16:57:11.120350 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:11 crc kubenswrapper[4954]: E1209 16:57:11.120490 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:11 crc kubenswrapper[4954]: E1209 16:57:11.120709 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.168685 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.168748 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.168763 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.168792 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.168810 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:11Z","lastTransitionTime":"2025-12-09T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.272281 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.272358 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.272369 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.272390 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.272403 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:11Z","lastTransitionTime":"2025-12-09T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.341343 4954 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.375170 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.375223 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.375234 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.375257 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.375275 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:11Z","lastTransitionTime":"2025-12-09T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.478809 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.478901 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.478977 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.479020 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.479043 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:11Z","lastTransitionTime":"2025-12-09T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.582422 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.582511 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.582528 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.582555 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.582573 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:11Z","lastTransitionTime":"2025-12-09T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.684842 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.684884 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.684895 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.684911 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.684921 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:11Z","lastTransitionTime":"2025-12-09T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.787700 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.787728 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.787736 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.787750 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.787759 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:11Z","lastTransitionTime":"2025-12-09T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.889672 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.889701 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.889709 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.889722 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.889729 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:11Z","lastTransitionTime":"2025-12-09T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.992201 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.992232 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.992240 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.992255 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:11 crc kubenswrapper[4954]: I1209 16:57:11.992264 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:11Z","lastTransitionTime":"2025-12-09T16:57:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.095182 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.095219 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.095228 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.095241 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.095250 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:12Z","lastTransitionTime":"2025-12-09T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.197497 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.197536 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.197548 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.197564 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.197577 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:12Z","lastTransitionTime":"2025-12-09T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.240855 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.300374 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.300406 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.300414 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.300430 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.300439 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:12Z","lastTransitionTime":"2025-12-09T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.402761 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.402808 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.402819 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.402835 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.402847 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:12Z","lastTransitionTime":"2025-12-09T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.505250 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.505535 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.505545 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.505562 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.505572 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:12Z","lastTransitionTime":"2025-12-09T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.608409 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.608477 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.608490 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.608519 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.608536 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:12Z","lastTransitionTime":"2025-12-09T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.711628 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.711670 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.711684 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.711708 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.711718 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:12Z","lastTransitionTime":"2025-12-09T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.814869 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.814923 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.814935 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.814957 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.814971 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:12Z","lastTransitionTime":"2025-12-09T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.917789 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.917838 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.917848 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.917868 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:12 crc kubenswrapper[4954]: I1209 16:57:12.917880 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:12Z","lastTransitionTime":"2025-12-09T16:57:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.020413 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.020457 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.020467 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.020483 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.020494 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:13Z","lastTransitionTime":"2025-12-09T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.119223 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:13 crc kubenswrapper[4954]: E1209 16:57:13.119351 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.119388 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:13 crc kubenswrapper[4954]: E1209 16:57:13.119860 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.120016 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:13 crc kubenswrapper[4954]: E1209 16:57:13.120177 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.123525 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.123554 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.123563 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.123580 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.123605 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:13Z","lastTransitionTime":"2025-12-09T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.135991 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.153635 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.170860 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.185560 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.203728 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.225123 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.226092 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.226124 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.226136 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.226160 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.226173 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:13Z","lastTransitionTime":"2025-12-09T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.241583 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.254250 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.266338 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.280096 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.295378 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.310255 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.328562 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.328634 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.328648 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.328670 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.328684 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:13Z","lastTransitionTime":"2025-12-09T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.329466 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.347969 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/0.log" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.350462 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65" exitCode=1 Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.350515 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.351092 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.351343 4954 scope.go:117] "RemoveContainer" containerID="4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.369004 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.383523 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.396283 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.411511 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.422994 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.431478 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.431519 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.431529 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.431546 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.431559 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:13Z","lastTransitionTime":"2025-12-09T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.436069 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.458936 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:12Z\\\",\\\"message\\\":\\\":208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 16:57:12.732670 6245 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 16:57:12.732967 6245 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 16:57:12.733014 6245 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 16:57:12.733021 6245 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 16:57:12.733035 6245 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 16:57:12.733040 6245 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 16:57:12.733055 6245 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 16:57:12.733071 6245 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 16:57:12.733079 6245 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 16:57:12.733098 6245 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1209 16:57:12.733102 6245 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 16:57:12.733136 6245 factory.go:656] Stopping watch factory\\\\nI1209 16:57:12.733140 6245 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 16:57:12.733167 6245 ovnkube.go:599] Stopped ovnkube\\\\nI1209 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.478197 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.490775 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.501952 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.514626 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.525084 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.534356 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.534407 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.534423 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.534442 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.534456 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:13Z","lastTransitionTime":"2025-12-09T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.538151 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.551738 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.563415 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.574740 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.588238 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:13Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.637089 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.637146 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.637157 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.637179 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.637191 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:13Z","lastTransitionTime":"2025-12-09T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.739823 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.739872 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.739884 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.739904 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.739917 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:13Z","lastTransitionTime":"2025-12-09T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.842874 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.842937 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.842955 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.842983 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.842998 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:13Z","lastTransitionTime":"2025-12-09T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.945999 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.946042 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.946052 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.946073 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:13 crc kubenswrapper[4954]: I1209 16:57:13.946086 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:13Z","lastTransitionTime":"2025-12-09T16:57:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.048931 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.048982 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.048995 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.049013 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.049023 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:14Z","lastTransitionTime":"2025-12-09T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.138506 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.151429 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.151519 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.151542 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.151552 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.151568 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.151578 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:14Z","lastTransitionTime":"2025-12-09T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.163398 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.189166 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.203834 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.219385 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.236637 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.254163 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.254212 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.254225 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.254246 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.254259 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:14Z","lastTransitionTime":"2025-12-09T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.265196 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:12Z\\\",\\\"message\\\":\\\":208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 16:57:12.732670 6245 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 16:57:12.732967 6245 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 16:57:12.733014 6245 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 16:57:12.733021 6245 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 16:57:12.733035 6245 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 16:57:12.733040 6245 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 16:57:12.733055 6245 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 16:57:12.733071 6245 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 16:57:12.733079 6245 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 16:57:12.733098 6245 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1209 16:57:12.733102 6245 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 16:57:12.733136 6245 factory.go:656] Stopping watch factory\\\\nI1209 16:57:12.733140 6245 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 16:57:12.733167 6245 ovnkube.go:599] Stopped ovnkube\\\\nI1209 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.280410 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.297029 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.317863 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.332675 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.349397 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.355865 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.355944 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.355960 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.355983 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.355997 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:14Z","lastTransitionTime":"2025-12-09T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.358561 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/0.log" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.361792 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.362464 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.368822 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.386215 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.407029 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.421446 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.453435 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.458991 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.459038 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.459052 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.459076 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.459092 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:14Z","lastTransitionTime":"2025-12-09T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.472618 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.495046 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:12Z\\\",\\\"message\\\":\\\":208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 16:57:12.732670 6245 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 16:57:12.732967 6245 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 16:57:12.733014 6245 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 16:57:12.733021 6245 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 16:57:12.733035 6245 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 16:57:12.733040 6245 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 16:57:12.733055 6245 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 16:57:12.733071 6245 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 16:57:12.733079 6245 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 16:57:12.733098 6245 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1209 16:57:12.733102 6245 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 16:57:12.733136 6245 factory.go:656] Stopping watch factory\\\\nI1209 16:57:12.733140 6245 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 16:57:12.733167 6245 ovnkube.go:599] Stopped ovnkube\\\\nI1209 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.515393 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.529094 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.546709 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.562100 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.562180 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.562203 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.562233 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.562226 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.562253 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:14Z","lastTransitionTime":"2025-12-09T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.581693 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.596845 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.616058 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.632061 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.646543 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.659974 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:14Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.665475 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.665535 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.665557 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.665678 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.665759 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:14Z","lastTransitionTime":"2025-12-09T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.768945 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.769014 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.769032 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.769060 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.769078 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:14Z","lastTransitionTime":"2025-12-09T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.872522 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.872581 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.872616 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.872636 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.872648 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:14Z","lastTransitionTime":"2025-12-09T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.975984 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.976059 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.976078 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.976107 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:14 crc kubenswrapper[4954]: I1209 16:57:14.976130 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:14Z","lastTransitionTime":"2025-12-09T16:57:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.023937 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6"] Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.024494 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.029409 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.032996 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.043127 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.058045 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.071953 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.078518 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.078577 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.078610 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.078632 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.078646 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.095287 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:12Z\\\",\\\"message\\\":\\\":208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 16:57:12.732670 6245 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 16:57:12.732967 6245 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 16:57:12.733014 6245 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 16:57:12.733021 6245 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 16:57:12.733035 6245 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 16:57:12.733040 6245 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 16:57:12.733055 6245 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 16:57:12.733071 6245 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 16:57:12.733079 6245 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 16:57:12.733098 6245 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1209 16:57:12.733102 6245 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 16:57:12.733136 6245 factory.go:656] Stopping watch factory\\\\nI1209 16:57:12.733140 6245 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 16:57:12.733167 6245 ovnkube.go:599] Stopped ovnkube\\\\nI1209 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.119863 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.119943 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.119913 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:15 crc kubenswrapper[4954]: E1209 16:57:15.120182 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.120112 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: E1209 16:57:15.120282 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:15 crc kubenswrapper[4954]: E1209 16:57:15.120389 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.137270 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.147860 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.148330 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/70175b95-63b3-485c-87a5-2e67167a7948-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.148364 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/70175b95-63b3-485c-87a5-2e67167a7948-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.148396 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrpkg\" (UniqueName: \"kubernetes.io/projected/70175b95-63b3-485c-87a5-2e67167a7948-kube-api-access-nrpkg\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.148452 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/70175b95-63b3-485c-87a5-2e67167a7948-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.158314 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.169078 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.181173 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.181224 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.181235 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.181252 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.181263 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.181442 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.195994 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.211414 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.226999 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.240749 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.250222 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/70175b95-63b3-485c-87a5-2e67167a7948-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.250325 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/70175b95-63b3-485c-87a5-2e67167a7948-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.250391 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrpkg\" (UniqueName: \"kubernetes.io/projected/70175b95-63b3-485c-87a5-2e67167a7948-kube-api-access-nrpkg\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.250470 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/70175b95-63b3-485c-87a5-2e67167a7948-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.250864 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/70175b95-63b3-485c-87a5-2e67167a7948-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.251451 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/70175b95-63b3-485c-87a5-2e67167a7948-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.255949 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/70175b95-63b3-485c-87a5-2e67167a7948-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.257008 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.270357 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrpkg\" (UniqueName: \"kubernetes.io/projected/70175b95-63b3-485c-87a5-2e67167a7948-kube-api-access-nrpkg\") pod \"ovnkube-control-plane-749d76644c-wxkr6\" (UID: \"70175b95-63b3-485c-87a5-2e67167a7948\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.280086 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.284550 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.284631 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.284650 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.284671 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.284683 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.345002 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" Dec 09 16:57:15 crc kubenswrapper[4954]: W1209 16:57:15.363945 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70175b95_63b3_485c_87a5_2e67167a7948.slice/crio-edca47d3e4d9bafca2b2fad8e3047924d568eaa54c659020256a8fc9d17de037 WatchSource:0}: Error finding container edca47d3e4d9bafca2b2fad8e3047924d568eaa54c659020256a8fc9d17de037: Status 404 returned error can't find the container with id edca47d3e4d9bafca2b2fad8e3047924d568eaa54c659020256a8fc9d17de037 Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.369329 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/1.log" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.370552 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/0.log" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.374265 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed" exitCode=1 Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.374327 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed"} Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.374383 4954 scope.go:117] "RemoveContainer" containerID="4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.375318 4954 scope.go:117] "RemoveContainer" containerID="780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed" Dec 09 16:57:15 crc kubenswrapper[4954]: E1209 16:57:15.375521 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.389739 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.389799 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.389815 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.389843 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.389860 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.392281 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.408220 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.432948 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e360242030a6b6bb7b9da63682fbeb4ccddd348ba3201fc0715b3d772c95a65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:12Z\\\",\\\"message\\\":\\\":208] Removed *v1.EgressFirewall event handler 9\\\\nI1209 16:57:12.732670 6245 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 16:57:12.732967 6245 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 16:57:12.733014 6245 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 16:57:12.733021 6245 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 16:57:12.733035 6245 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 16:57:12.733040 6245 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 16:57:12.733055 6245 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1209 16:57:12.733071 6245 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1209 16:57:12.733079 6245 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 16:57:12.733098 6245 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1209 16:57:12.733102 6245 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1209 16:57:12.733136 6245 factory.go:656] Stopping watch factory\\\\nI1209 16:57:12.733140 6245 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1209 16:57:12.733167 6245 ovnkube.go:599] Stopped ovnkube\\\\nI1209 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:14Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 16:57:14.445398 6363 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:14.445449 6363 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-met\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.459273 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.478334 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.493081 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.493120 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.493131 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.493148 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.493158 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.500829 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.515746 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.529036 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.544261 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.561717 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.579137 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.594653 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.596732 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.596787 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.596802 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.596824 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.596839 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.608169 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.619882 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.639755 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.657203 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.699513 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.699569 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.699585 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.699633 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.699650 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.802652 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.802762 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.802774 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.802798 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.802812 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.844675 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.844726 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.844738 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.844755 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.844766 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: E1209 16:57:15.856573 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.860887 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.860941 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.860953 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.860980 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.860993 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: E1209 16:57:15.877156 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.882131 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.882183 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.882197 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.882219 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.882233 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: E1209 16:57:15.897988 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.902412 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.902445 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.902457 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.902479 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.902493 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: E1209 16:57:15.917984 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.922004 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.922111 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.922130 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.922147 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.922160 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:15 crc kubenswrapper[4954]: E1209 16:57:15.936233 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:15Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:15 crc kubenswrapper[4954]: E1209 16:57:15.936384 4954 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.937991 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.938039 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.938051 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.938080 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:15 crc kubenswrapper[4954]: I1209 16:57:15.938093 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:15Z","lastTransitionTime":"2025-12-09T16:57:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.041349 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.041426 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.041448 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.041479 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.041497 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:16Z","lastTransitionTime":"2025-12-09T16:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.144359 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.144407 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.144425 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.144442 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.144453 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:16Z","lastTransitionTime":"2025-12-09T16:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.247293 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.247329 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.247337 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.247352 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.247372 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:16Z","lastTransitionTime":"2025-12-09T16:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.350833 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.350900 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.350917 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.350940 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.350954 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:16Z","lastTransitionTime":"2025-12-09T16:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.383979 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/1.log" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.389817 4954 scope.go:117] "RemoveContainer" containerID="780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed" Dec 09 16:57:16 crc kubenswrapper[4954]: E1209 16:57:16.390171 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.390678 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" event={"ID":"70175b95-63b3-485c-87a5-2e67167a7948","Type":"ContainerStarted","Data":"7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.390772 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" event={"ID":"70175b95-63b3-485c-87a5-2e67167a7948","Type":"ContainerStarted","Data":"e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.390806 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" event={"ID":"70175b95-63b3-485c-87a5-2e67167a7948","Type":"ContainerStarted","Data":"edca47d3e4d9bafca2b2fad8e3047924d568eaa54c659020256a8fc9d17de037"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.402177 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.411864 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.421914 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.433278 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.446322 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.452718 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.452763 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.452798 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.452819 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.452831 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:16Z","lastTransitionTime":"2025-12-09T16:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.459584 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.470431 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.485247 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.490284 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-ffkzn"] Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.490780 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:16 crc kubenswrapper[4954]: E1209 16:57:16.490851 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.497910 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.509554 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.520524 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.538616 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.555430 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.555931 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.556001 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.556016 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.556034 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.556049 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:16Z","lastTransitionTime":"2025-12-09T16:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.567957 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.580406 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.597305 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:14Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 16:57:14.445398 6363 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:14.445449 6363 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-met\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.607362 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.617931 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.629818 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.639829 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.649999 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.658444 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.658484 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.658494 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.658509 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.658517 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:16Z","lastTransitionTime":"2025-12-09T16:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.661643 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.670009 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.670040 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf6cb\" (UniqueName: \"kubernetes.io/projected/856fc314-24bd-403e-a4b0-3dcd73eba595-kube-api-access-lf6cb\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.673032 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.686146 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.697952 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.709379 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.718011 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.738545 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:14Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 16:57:14.445398 6363 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:14.445449 6363 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-met\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.748563 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.760262 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.760294 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.760306 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.760323 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.760336 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:16Z","lastTransitionTime":"2025-12-09T16:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.768290 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.770622 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.770661 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf6cb\" (UniqueName: \"kubernetes.io/projected/856fc314-24bd-403e-a4b0-3dcd73eba595-kube-api-access-lf6cb\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:16 crc kubenswrapper[4954]: E1209 16:57:16.770789 4954 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:16 crc kubenswrapper[4954]: E1209 16:57:16.770856 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs podName:856fc314-24bd-403e-a4b0-3dcd73eba595 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:17.270838144 +0000 UTC m=+33.659011964 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs") pod "network-metrics-daemon-ffkzn" (UID: "856fc314-24bd-403e-a4b0-3dcd73eba595") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.781413 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.787742 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf6cb\" (UniqueName: \"kubernetes.io/projected/856fc314-24bd-403e-a4b0-3dcd73eba595-kube-api-access-lf6cb\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.793457 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.811991 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:16Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.863308 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.863350 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.863360 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.863376 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.863388 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:16Z","lastTransitionTime":"2025-12-09T16:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.965980 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.966030 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.966043 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.966062 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:16 crc kubenswrapper[4954]: I1209 16:57:16.966074 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:16Z","lastTransitionTime":"2025-12-09T16:57:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.068350 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.068381 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.068394 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.068409 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.068420 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:17Z","lastTransitionTime":"2025-12-09T16:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.119770 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.119830 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.119862 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:17 crc kubenswrapper[4954]: E1209 16:57:17.119894 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:17 crc kubenswrapper[4954]: E1209 16:57:17.119985 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:17 crc kubenswrapper[4954]: E1209 16:57:17.120044 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.170777 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.170816 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.170825 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.170840 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.170849 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:17Z","lastTransitionTime":"2025-12-09T16:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.273632 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.273666 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.273674 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.273687 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.273697 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:17Z","lastTransitionTime":"2025-12-09T16:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.275235 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:17 crc kubenswrapper[4954]: E1209 16:57:17.275330 4954 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:17 crc kubenswrapper[4954]: E1209 16:57:17.275381 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs podName:856fc314-24bd-403e-a4b0-3dcd73eba595 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:18.275367887 +0000 UTC m=+34.663541707 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs") pod "network-metrics-daemon-ffkzn" (UID: "856fc314-24bd-403e-a4b0-3dcd73eba595") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.375526 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.375560 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.375570 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.375586 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.375613 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:17Z","lastTransitionTime":"2025-12-09T16:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.478524 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.478556 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.478564 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.478579 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.478588 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:17Z","lastTransitionTime":"2025-12-09T16:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.580478 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.580516 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.580527 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.580542 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.580552 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:17Z","lastTransitionTime":"2025-12-09T16:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.682912 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.683167 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.683244 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.683306 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.683372 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:17Z","lastTransitionTime":"2025-12-09T16:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.785808 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.785851 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.785862 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.785880 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.785892 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:17Z","lastTransitionTime":"2025-12-09T16:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.888831 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.888871 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.888879 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.888892 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.888901 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:17Z","lastTransitionTime":"2025-12-09T16:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.991427 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.991471 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.991488 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.991515 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:17 crc kubenswrapper[4954]: I1209 16:57:17.991527 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:17Z","lastTransitionTime":"2025-12-09T16:57:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.094303 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.094337 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.094347 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.094359 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.094368 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:18Z","lastTransitionTime":"2025-12-09T16:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.120118 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.120265 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.196359 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.196619 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.196691 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.196762 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.196831 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:18Z","lastTransitionTime":"2025-12-09T16:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.285056 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.285281 4954 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.285422 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs podName:856fc314-24bd-403e-a4b0-3dcd73eba595 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:20.285386206 +0000 UTC m=+36.673560066 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs") pod "network-metrics-daemon-ffkzn" (UID: "856fc314-24bd-403e-a4b0-3dcd73eba595") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.298777 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.298842 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.298864 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.298891 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.298908 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:18Z","lastTransitionTime":"2025-12-09T16:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.401416 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.401461 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.401472 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.401489 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.401500 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:18Z","lastTransitionTime":"2025-12-09T16:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.504388 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.504441 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.504457 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.504478 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.504493 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:18Z","lastTransitionTime":"2025-12-09T16:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.607253 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.607325 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.607351 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.607383 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.607410 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:18Z","lastTransitionTime":"2025-12-09T16:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.710557 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.710671 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.710696 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.710727 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.710744 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:18Z","lastTransitionTime":"2025-12-09T16:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.813508 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.813556 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.813566 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.813579 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.813607 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:18Z","lastTransitionTime":"2025-12-09T16:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.892073 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.892288 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:57:34.892249971 +0000 UTC m=+51.280423831 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.916456 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.916524 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.916541 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.916569 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.916587 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:18Z","lastTransitionTime":"2025-12-09T16:57:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.993310 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.993420 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.993454 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:18 crc kubenswrapper[4954]: I1209 16:57:18.993495 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993547 4954 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993651 4954 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993660 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993704 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993721 4954 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993659 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:34.9936371 +0000 UTC m=+51.381810920 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993660 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993808 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:34.993786317 +0000 UTC m=+51.381960207 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993838 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:34.993825959 +0000 UTC m=+51.381999869 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993843 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993863 4954 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:18 crc kubenswrapper[4954]: E1209 16:57:18.993913 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:34.993891441 +0000 UTC m=+51.382065271 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.019675 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.019758 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.019802 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.019822 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.019834 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:19Z","lastTransitionTime":"2025-12-09T16:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.119719 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.119778 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.119861 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:19 crc kubenswrapper[4954]: E1209 16:57:19.119896 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:19 crc kubenswrapper[4954]: E1209 16:57:19.119996 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:19 crc kubenswrapper[4954]: E1209 16:57:19.120122 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.121640 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.121677 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.121692 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.121711 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.121726 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:19Z","lastTransitionTime":"2025-12-09T16:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.225342 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.225413 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.225433 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.225462 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.225481 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:19Z","lastTransitionTime":"2025-12-09T16:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.328256 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.328322 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.328345 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.328374 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.328396 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:19Z","lastTransitionTime":"2025-12-09T16:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.430519 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.430555 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.430566 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.430580 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.430611 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:19Z","lastTransitionTime":"2025-12-09T16:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.533456 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.533498 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.533507 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.533519 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.533528 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:19Z","lastTransitionTime":"2025-12-09T16:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.636473 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.636516 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.636530 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.636545 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.636557 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:19Z","lastTransitionTime":"2025-12-09T16:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.740133 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.740178 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.740189 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.740205 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.740214 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:19Z","lastTransitionTime":"2025-12-09T16:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.842498 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.842544 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.842562 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.842578 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.842588 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:19Z","lastTransitionTime":"2025-12-09T16:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.947114 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.947210 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.947236 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.947266 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:19 crc kubenswrapper[4954]: I1209 16:57:19.947292 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:19Z","lastTransitionTime":"2025-12-09T16:57:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.051125 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.051187 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.051195 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.051211 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.051221 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:20Z","lastTransitionTime":"2025-12-09T16:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.119396 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:20 crc kubenswrapper[4954]: E1209 16:57:20.119632 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.154677 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.154715 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.154727 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.154744 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.154755 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:20Z","lastTransitionTime":"2025-12-09T16:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.257507 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.257549 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.257560 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.257578 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.257606 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:20Z","lastTransitionTime":"2025-12-09T16:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.308210 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:20 crc kubenswrapper[4954]: E1209 16:57:20.308365 4954 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:20 crc kubenswrapper[4954]: E1209 16:57:20.308434 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs podName:856fc314-24bd-403e-a4b0-3dcd73eba595 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:24.308416113 +0000 UTC m=+40.696589943 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs") pod "network-metrics-daemon-ffkzn" (UID: "856fc314-24bd-403e-a4b0-3dcd73eba595") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.360112 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.360146 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.360159 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.360174 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.360186 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:20Z","lastTransitionTime":"2025-12-09T16:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.462751 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.462806 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.462815 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.462828 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.462840 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:20Z","lastTransitionTime":"2025-12-09T16:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.565704 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.565767 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.565783 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.565807 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.565824 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:20Z","lastTransitionTime":"2025-12-09T16:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.669150 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.669203 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.669219 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.669247 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.669265 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:20Z","lastTransitionTime":"2025-12-09T16:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.771409 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.771463 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.771471 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.771486 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.771494 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:20Z","lastTransitionTime":"2025-12-09T16:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.874305 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.874342 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.874353 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.874389 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.874399 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:20Z","lastTransitionTime":"2025-12-09T16:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.978474 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.978551 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.978566 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.978584 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:20 crc kubenswrapper[4954]: I1209 16:57:20.978662 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:20Z","lastTransitionTime":"2025-12-09T16:57:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.082486 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.082565 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.082582 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.082627 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.082641 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:21Z","lastTransitionTime":"2025-12-09T16:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.119097 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.119142 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.119114 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:21 crc kubenswrapper[4954]: E1209 16:57:21.119241 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:21 crc kubenswrapper[4954]: E1209 16:57:21.119404 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:21 crc kubenswrapper[4954]: E1209 16:57:21.119507 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.184638 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.184683 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.184692 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.184706 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.184717 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:21Z","lastTransitionTime":"2025-12-09T16:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.287016 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.287059 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.287075 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.287091 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.287102 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:21Z","lastTransitionTime":"2025-12-09T16:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.389643 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.389701 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.389722 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.389755 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.389779 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:21Z","lastTransitionTime":"2025-12-09T16:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.492689 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.492734 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.492742 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.492759 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.492768 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:21Z","lastTransitionTime":"2025-12-09T16:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.594585 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.594669 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.594688 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.594712 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.594729 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:21Z","lastTransitionTime":"2025-12-09T16:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.697761 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.697830 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.697851 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.697879 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.697900 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:21Z","lastTransitionTime":"2025-12-09T16:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.799905 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.799944 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.799953 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.799967 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.799977 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:21Z","lastTransitionTime":"2025-12-09T16:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.901549 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.901610 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.901624 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.901640 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:21 crc kubenswrapper[4954]: I1209 16:57:21.901649 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:21Z","lastTransitionTime":"2025-12-09T16:57:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.003977 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.004029 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.004041 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.004056 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.004066 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:22Z","lastTransitionTime":"2025-12-09T16:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.106982 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.107063 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.107084 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.107138 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.107161 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:22Z","lastTransitionTime":"2025-12-09T16:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.119390 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:22 crc kubenswrapper[4954]: E1209 16:57:22.119569 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.209384 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.209431 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.209444 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.209461 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.209475 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:22Z","lastTransitionTime":"2025-12-09T16:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.312193 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.312619 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.312633 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.312653 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.312664 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:22Z","lastTransitionTime":"2025-12-09T16:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.415225 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.415265 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.415280 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.415294 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.415304 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:22Z","lastTransitionTime":"2025-12-09T16:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.517548 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.517615 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.517627 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.517644 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.517658 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:22Z","lastTransitionTime":"2025-12-09T16:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.619918 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.619956 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.619965 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.619978 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.619987 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:22Z","lastTransitionTime":"2025-12-09T16:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.721866 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.721912 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.721924 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.721939 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.721951 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:22Z","lastTransitionTime":"2025-12-09T16:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.824178 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.824221 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.824230 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.824244 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.824254 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:22Z","lastTransitionTime":"2025-12-09T16:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.926432 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.926470 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.926478 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.926495 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:22 crc kubenswrapper[4954]: I1209 16:57:22.926503 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:22Z","lastTransitionTime":"2025-12-09T16:57:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.028872 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.028913 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.028921 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.028937 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.028947 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:23Z","lastTransitionTime":"2025-12-09T16:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.119173 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.119247 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.119246 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:23 crc kubenswrapper[4954]: E1209 16:57:23.119327 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:23 crc kubenswrapper[4954]: E1209 16:57:23.119420 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:23 crc kubenswrapper[4954]: E1209 16:57:23.119510 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.135739 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.135788 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.135798 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.135813 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.135823 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:23Z","lastTransitionTime":"2025-12-09T16:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.237685 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.238139 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.238215 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.238274 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.238371 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:23Z","lastTransitionTime":"2025-12-09T16:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.340923 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.340972 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.340988 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.341007 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.341022 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:23Z","lastTransitionTime":"2025-12-09T16:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.443325 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.443378 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.443394 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.443410 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.443423 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:23Z","lastTransitionTime":"2025-12-09T16:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.546619 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.546708 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.546725 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.546744 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.546756 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:23Z","lastTransitionTime":"2025-12-09T16:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.649349 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.649400 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.649415 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.649434 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.649452 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:23Z","lastTransitionTime":"2025-12-09T16:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.751438 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.751477 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.751486 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.751499 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.751510 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:23Z","lastTransitionTime":"2025-12-09T16:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.854508 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.854654 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.854682 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.854721 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.854765 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:23Z","lastTransitionTime":"2025-12-09T16:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.957093 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.957167 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.957180 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.957200 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:23 crc kubenswrapper[4954]: I1209 16:57:23.957212 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:23Z","lastTransitionTime":"2025-12-09T16:57:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.059756 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.059797 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.059806 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.059827 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.059840 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:24Z","lastTransitionTime":"2025-12-09T16:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.119714 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:24 crc kubenswrapper[4954]: E1209 16:57:24.119916 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.138648 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.151740 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.163369 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.163483 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.163547 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.163589 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.163690 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:24Z","lastTransitionTime":"2025-12-09T16:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.166406 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.180563 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.194970 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.211701 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.226534 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.244234 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.263324 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.271884 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.271963 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.271984 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.272016 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.272040 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:24Z","lastTransitionTime":"2025-12-09T16:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.284632 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.297226 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.317079 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.346250 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:14Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 16:57:14.445398 6363 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:14.445449 6363 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-met\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.349839 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:24 crc kubenswrapper[4954]: E1209 16:57:24.350096 4954 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:24 crc kubenswrapper[4954]: E1209 16:57:24.350231 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs podName:856fc314-24bd-403e-a4b0-3dcd73eba595 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:32.350198982 +0000 UTC m=+48.738373002 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs") pod "network-metrics-daemon-ffkzn" (UID: "856fc314-24bd-403e-a4b0-3dcd73eba595") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.359217 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.374516 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.374553 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.374564 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.374581 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.374610 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:24Z","lastTransitionTime":"2025-12-09T16:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.381387 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.395183 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.406412 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:24Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.478119 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.478181 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.478199 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.478229 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.478253 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:24Z","lastTransitionTime":"2025-12-09T16:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.581876 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.581950 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.581964 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.581990 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.582002 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:24Z","lastTransitionTime":"2025-12-09T16:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.685772 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.685852 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.685872 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.685903 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.685923 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:24Z","lastTransitionTime":"2025-12-09T16:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.789518 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.789615 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.789633 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.789659 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.789675 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:24Z","lastTransitionTime":"2025-12-09T16:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.892725 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.892806 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.892829 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.892924 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.892949 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:24Z","lastTransitionTime":"2025-12-09T16:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.995900 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.995955 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.995967 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.995991 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:24 crc kubenswrapper[4954]: I1209 16:57:24.996008 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:24Z","lastTransitionTime":"2025-12-09T16:57:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.099625 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.099688 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.099703 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.099736 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.099751 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:25Z","lastTransitionTime":"2025-12-09T16:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.119510 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.119576 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.119700 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:25 crc kubenswrapper[4954]: E1209 16:57:25.119749 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:25 crc kubenswrapper[4954]: E1209 16:57:25.119972 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:25 crc kubenswrapper[4954]: E1209 16:57:25.120204 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.204232 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.204359 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.204384 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.204419 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.204441 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:25Z","lastTransitionTime":"2025-12-09T16:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.307198 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.307244 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.307255 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.307276 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.307289 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:25Z","lastTransitionTime":"2025-12-09T16:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.409118 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.409152 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.409161 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.409176 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.409186 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:25Z","lastTransitionTime":"2025-12-09T16:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.512753 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.513447 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.513480 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.513521 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.513539 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:25Z","lastTransitionTime":"2025-12-09T16:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.616675 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.616752 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.616775 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.616822 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.616837 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:25Z","lastTransitionTime":"2025-12-09T16:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.720467 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.720555 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.720585 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.720669 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.720693 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:25Z","lastTransitionTime":"2025-12-09T16:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.823607 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.823656 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.823666 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.823681 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.823692 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:25Z","lastTransitionTime":"2025-12-09T16:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.926670 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.926729 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.926755 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.926781 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:25 crc kubenswrapper[4954]: I1209 16:57:25.926798 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:25Z","lastTransitionTime":"2025-12-09T16:57:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.029328 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.029365 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.029374 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.029388 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.029397 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.068811 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.068851 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.068860 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.068874 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.068886 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: E1209 16:57:26.084722 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:26Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.089234 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.089274 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.089284 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.089299 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.089308 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: E1209 16:57:26.108136 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:26Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.112063 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.112101 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.112110 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.112126 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.112136 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.119522 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:26 crc kubenswrapper[4954]: E1209 16:57:26.119665 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:26 crc kubenswrapper[4954]: E1209 16:57:26.125090 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:26Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.129760 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.129813 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.129850 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.129870 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.129882 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: E1209 16:57:26.144315 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:26Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.148457 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.148499 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.148511 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.148529 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.148540 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: E1209 16:57:26.166279 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:26Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:26 crc kubenswrapper[4954]: E1209 16:57:26.166447 4954 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.168265 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.168921 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.168947 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.168966 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.168979 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.271662 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.271715 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.271732 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.271758 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.271776 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.374390 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.374444 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.374452 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.374467 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.374475 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.476286 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.476326 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.476339 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.476354 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.476367 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.578390 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.578431 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.578442 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.578464 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.578475 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.681429 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.681491 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.681505 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.681533 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.681552 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.783530 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.783576 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.783584 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.783621 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.783634 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.887296 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.887375 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.887401 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.887435 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.887458 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.990666 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.990750 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.990771 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.990798 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:26 crc kubenswrapper[4954]: I1209 16:57:26.990813 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:26Z","lastTransitionTime":"2025-12-09T16:57:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.094442 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.094501 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.094515 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.094533 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.094543 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:27Z","lastTransitionTime":"2025-12-09T16:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.119949 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:27 crc kubenswrapper[4954]: E1209 16:57:27.120117 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.120566 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.120640 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:27 crc kubenswrapper[4954]: E1209 16:57:27.120729 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:27 crc kubenswrapper[4954]: E1209 16:57:27.120795 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.121403 4954 scope.go:117] "RemoveContainer" containerID="780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.197524 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.197578 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.197613 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.197641 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.197660 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:27Z","lastTransitionTime":"2025-12-09T16:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.301629 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.301692 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.301704 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.301725 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.301737 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:27Z","lastTransitionTime":"2025-12-09T16:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.404914 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.404978 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.405007 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.405032 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.405050 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:27Z","lastTransitionTime":"2025-12-09T16:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.427748 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/1.log" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.430881 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7"} Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.431511 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.446402 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.464976 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.481938 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.497233 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.508056 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.508100 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.508110 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.508130 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.508144 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:27Z","lastTransitionTime":"2025-12-09T16:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.514694 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.528661 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.542553 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.569268 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.582081 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.595949 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.607102 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.611665 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.611697 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.611708 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.611724 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.611733 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:27Z","lastTransitionTime":"2025-12-09T16:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.623699 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:14Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 16:57:14.445398 6363 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:14.445449 6363 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-met\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.638014 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.649531 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.694287 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.705379 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.714164 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.714195 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.714215 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.714230 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.714238 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:27Z","lastTransitionTime":"2025-12-09T16:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.716173 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:27Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.816250 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.816290 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.816298 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.816311 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.816320 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:27Z","lastTransitionTime":"2025-12-09T16:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.918278 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.918310 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.918319 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.918334 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:27 crc kubenswrapper[4954]: I1209 16:57:27.918342 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:27Z","lastTransitionTime":"2025-12-09T16:57:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.020254 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.020301 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.020316 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.020335 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.020345 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:28Z","lastTransitionTime":"2025-12-09T16:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.119199 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:28 crc kubenswrapper[4954]: E1209 16:57:28.119327 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.122964 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.123013 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.123030 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.123051 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.123066 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:28Z","lastTransitionTime":"2025-12-09T16:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.225353 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.225401 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.225414 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.225432 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.225458 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:28Z","lastTransitionTime":"2025-12-09T16:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.327818 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.327878 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.327889 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.327906 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.327917 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:28Z","lastTransitionTime":"2025-12-09T16:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.431781 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.431840 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.431851 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.431868 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.431878 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:28Z","lastTransitionTime":"2025-12-09T16:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.435494 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/2.log" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.436420 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/1.log" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.439370 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7" exitCode=1 Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.439426 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.439476 4954 scope.go:117] "RemoveContainer" containerID="780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.440144 4954 scope.go:117] "RemoveContainer" containerID="c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7" Dec 09 16:57:28 crc kubenswrapper[4954]: E1209 16:57:28.440329 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.457014 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.473312 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.486491 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.499771 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.519195 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://780bc9c50f3897896b23b40945a9060e021404ff5f72c66ae7fb7d8e4acd65ed\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:14Z\\\",\\\"message\\\":\\\"rce:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 16:57:14.445398 6363 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:14.445449 6363 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-met\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:28Z\\\",\\\"message\\\":\\\"ce_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:27.963752 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1209 16:57:27.963753 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963755 6583 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-ffkzn in node crc\\\\nI1209 16:57:27.963761 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963765 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963771 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963776 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.530363 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.534826 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.534863 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.534871 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.534886 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.534898 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:28Z","lastTransitionTime":"2025-12-09T16:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.549665 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.561515 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.573248 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.585124 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.595446 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.606104 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.618010 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.634244 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.637268 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.637310 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.637332 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.637353 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.637366 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:28Z","lastTransitionTime":"2025-12-09T16:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.646998 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.660132 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.673986 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:28Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.741096 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.741168 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.741181 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.741199 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.741210 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:28Z","lastTransitionTime":"2025-12-09T16:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.843549 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.843616 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.843629 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.843644 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.843653 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:28Z","lastTransitionTime":"2025-12-09T16:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.946321 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.946399 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.946413 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.946438 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:28 crc kubenswrapper[4954]: I1209 16:57:28.946452 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:28Z","lastTransitionTime":"2025-12-09T16:57:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.049111 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.049168 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.049178 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.049196 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.049205 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:29Z","lastTransitionTime":"2025-12-09T16:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.119105 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.119196 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:29 crc kubenswrapper[4954]: E1209 16:57:29.119251 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.119326 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:29 crc kubenswrapper[4954]: E1209 16:57:29.119396 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:29 crc kubenswrapper[4954]: E1209 16:57:29.119483 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.152216 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.152276 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.152289 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.152308 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.152321 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:29Z","lastTransitionTime":"2025-12-09T16:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.254508 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.254555 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.254563 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.254583 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.254614 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:29Z","lastTransitionTime":"2025-12-09T16:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.358417 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.358470 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.358483 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.358504 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.358518 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:29Z","lastTransitionTime":"2025-12-09T16:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.446974 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/2.log" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.452987 4954 scope.go:117] "RemoveContainer" containerID="c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7" Dec 09 16:57:29 crc kubenswrapper[4954]: E1209 16:57:29.453156 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.461307 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.461365 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.461390 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.461422 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.461449 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:29Z","lastTransitionTime":"2025-12-09T16:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.471802 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.487850 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.502324 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.537990 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.555361 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.564815 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.564894 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.564918 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.564942 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.564956 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:29Z","lastTransitionTime":"2025-12-09T16:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.573548 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.590635 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.616444 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:28Z\\\",\\\"message\\\":\\\"ce_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:27.963752 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1209 16:57:27.963753 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963755 6583 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-ffkzn in node crc\\\\nI1209 16:57:27.963761 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963765 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963771 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963776 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.631015 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.648055 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.665505 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.667446 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.667492 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.667503 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.667522 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.667533 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:29Z","lastTransitionTime":"2025-12-09T16:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.680228 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.695249 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.714434 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.731177 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.750582 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.768528 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:29Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.770147 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.770173 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.770185 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.770204 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.770217 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:29Z","lastTransitionTime":"2025-12-09T16:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.873357 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.873403 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.873415 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.873435 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.873446 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:29Z","lastTransitionTime":"2025-12-09T16:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.976701 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.976742 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.976753 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.976770 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:29 crc kubenswrapper[4954]: I1209 16:57:29.976785 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:29Z","lastTransitionTime":"2025-12-09T16:57:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.080106 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.080216 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.080232 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.080260 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.080288 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:30Z","lastTransitionTime":"2025-12-09T16:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.120267 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:30 crc kubenswrapper[4954]: E1209 16:57:30.120656 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.184360 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.184456 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.184478 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.184510 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.184530 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:30Z","lastTransitionTime":"2025-12-09T16:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.288186 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.288303 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.288315 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.288335 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.288346 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:30Z","lastTransitionTime":"2025-12-09T16:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.392820 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.392913 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.392941 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.392971 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.392993 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:30Z","lastTransitionTime":"2025-12-09T16:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.497260 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.497342 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.497371 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.497410 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.497439 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:30Z","lastTransitionTime":"2025-12-09T16:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.600684 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.600791 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.600817 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.600850 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.600871 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:30Z","lastTransitionTime":"2025-12-09T16:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.703324 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.703379 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.703393 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.703410 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.703422 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:30Z","lastTransitionTime":"2025-12-09T16:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.806987 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.807093 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.807109 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.807132 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.807147 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:30Z","lastTransitionTime":"2025-12-09T16:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.910351 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.911099 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.911144 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.911174 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:30 crc kubenswrapper[4954]: I1209 16:57:30.911193 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:30Z","lastTransitionTime":"2025-12-09T16:57:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.014155 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.014197 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.014210 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.014225 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.014237 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:31Z","lastTransitionTime":"2025-12-09T16:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.117382 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.117414 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.117423 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.117438 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.117476 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:31Z","lastTransitionTime":"2025-12-09T16:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.119944 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.119956 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.119987 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:31 crc kubenswrapper[4954]: E1209 16:57:31.120043 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:31 crc kubenswrapper[4954]: E1209 16:57:31.120102 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:31 crc kubenswrapper[4954]: E1209 16:57:31.120168 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.220999 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.221059 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.221072 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.221098 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.221115 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:31Z","lastTransitionTime":"2025-12-09T16:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.324248 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.324482 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.324543 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.324629 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.324715 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:31Z","lastTransitionTime":"2025-12-09T16:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.427572 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.427702 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.427718 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.427734 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.427746 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:31Z","lastTransitionTime":"2025-12-09T16:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.530144 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.530171 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.530179 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.530193 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.530202 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:31Z","lastTransitionTime":"2025-12-09T16:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.632656 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.632709 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.632721 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.632737 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.632747 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:31Z","lastTransitionTime":"2025-12-09T16:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.735498 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.735537 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.735550 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.735568 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.735580 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:31Z","lastTransitionTime":"2025-12-09T16:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.838401 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.838448 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.838458 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.838475 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.838486 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:31Z","lastTransitionTime":"2025-12-09T16:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.941118 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.941158 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.941169 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.941189 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:31 crc kubenswrapper[4954]: I1209 16:57:31.941200 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:31Z","lastTransitionTime":"2025-12-09T16:57:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.044339 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.044453 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.044465 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.044485 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.044502 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:32Z","lastTransitionTime":"2025-12-09T16:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.119571 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:32 crc kubenswrapper[4954]: E1209 16:57:32.119755 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.147210 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.147261 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.147270 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.147286 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.147296 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:32Z","lastTransitionTime":"2025-12-09T16:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.250353 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.250386 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.250397 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.250411 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.250421 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:32Z","lastTransitionTime":"2025-12-09T16:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.352638 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.352689 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.352704 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.352723 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.352733 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:32Z","lastTransitionTime":"2025-12-09T16:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.440463 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:32 crc kubenswrapper[4954]: E1209 16:57:32.440653 4954 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:32 crc kubenswrapper[4954]: E1209 16:57:32.440780 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs podName:856fc314-24bd-403e-a4b0-3dcd73eba595 nodeName:}" failed. No retries permitted until 2025-12-09 16:57:48.440757853 +0000 UTC m=+64.828931673 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs") pod "network-metrics-daemon-ffkzn" (UID: "856fc314-24bd-403e-a4b0-3dcd73eba595") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.455353 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.455393 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.455405 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.455422 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.455433 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:32Z","lastTransitionTime":"2025-12-09T16:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.558117 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.558172 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.558185 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.558199 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.558209 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:32Z","lastTransitionTime":"2025-12-09T16:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.661071 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.661123 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.661133 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.661152 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.661164 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:32Z","lastTransitionTime":"2025-12-09T16:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.763770 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.764024 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.764040 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.764057 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.764066 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:32Z","lastTransitionTime":"2025-12-09T16:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.867190 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.867233 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.867243 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.867258 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.867269 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:32Z","lastTransitionTime":"2025-12-09T16:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.969969 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.970008 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.970016 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.970033 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:32 crc kubenswrapper[4954]: I1209 16:57:32.970042 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:32Z","lastTransitionTime":"2025-12-09T16:57:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.072503 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.072544 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.072553 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.072569 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.072579 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:33Z","lastTransitionTime":"2025-12-09T16:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.119175 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.119196 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.119257 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:33 crc kubenswrapper[4954]: E1209 16:57:33.119294 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:33 crc kubenswrapper[4954]: E1209 16:57:33.119401 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:33 crc kubenswrapper[4954]: E1209 16:57:33.119486 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.175430 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.175464 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.175474 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.175504 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.175513 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:33Z","lastTransitionTime":"2025-12-09T16:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.277954 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.277999 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.278012 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.278028 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.278040 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:33Z","lastTransitionTime":"2025-12-09T16:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.380630 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.380669 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.380678 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.380694 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.380710 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:33Z","lastTransitionTime":"2025-12-09T16:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.482957 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.482996 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.483021 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.483041 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.483051 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:33Z","lastTransitionTime":"2025-12-09T16:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.585303 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.585336 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.585345 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.585357 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.585367 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:33Z","lastTransitionTime":"2025-12-09T16:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.687323 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.687391 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.687403 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.687417 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.687428 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:33Z","lastTransitionTime":"2025-12-09T16:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.789898 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.789946 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.789956 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.789973 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.789985 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:33Z","lastTransitionTime":"2025-12-09T16:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.892724 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.892795 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.892807 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.892823 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.892833 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:33Z","lastTransitionTime":"2025-12-09T16:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.994584 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.994646 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.994656 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.994699 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:33 crc kubenswrapper[4954]: I1209 16:57:33.994716 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:33Z","lastTransitionTime":"2025-12-09T16:57:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.096809 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.096844 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.096854 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.096868 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.096880 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:34Z","lastTransitionTime":"2025-12-09T16:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.119361 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:34 crc kubenswrapper[4954]: E1209 16:57:34.119482 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.133574 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.145629 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.156381 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.176879 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.187339 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.198268 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.199096 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.199127 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.199139 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.199155 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.199166 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:34Z","lastTransitionTime":"2025-12-09T16:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.210484 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.227208 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:28Z\\\",\\\"message\\\":\\\"ce_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:27.963752 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1209 16:57:27.963753 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963755 6583 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-ffkzn in node crc\\\\nI1209 16:57:27.963761 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963765 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963771 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963776 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.236915 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.249636 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.259935 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.269921 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.280871 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.297484 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.301523 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.301578 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.301618 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.301643 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.301659 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:34Z","lastTransitionTime":"2025-12-09T16:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.312664 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.324824 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.340588 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.381409 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.390689 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.393000 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.404038 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.404073 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.404083 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.404098 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.404107 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:34Z","lastTransitionTime":"2025-12-09T16:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.406631 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.416896 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.427261 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.441932 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.453269 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.469107 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.483818 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.494824 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.503584 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.506664 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.506767 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.506838 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.506914 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.506970 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:34Z","lastTransitionTime":"2025-12-09T16:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.517503 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.531182 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.543982 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.559011 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.582463 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:28Z\\\",\\\"message\\\":\\\"ce_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:27.963752 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1209 16:57:27.963753 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963755 6583 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-ffkzn in node crc\\\\nI1209 16:57:27.963761 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963765 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963771 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963776 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.599799 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.614375 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.614705 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.614946 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.615161 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.615351 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:34Z","lastTransitionTime":"2025-12-09T16:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.630662 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:34Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.718238 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.718274 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.718286 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.718302 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.718312 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:34Z","lastTransitionTime":"2025-12-09T16:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.820695 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.820746 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.820755 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.820774 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.820789 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:34Z","lastTransitionTime":"2025-12-09T16:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.923312 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.923353 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.923367 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.923383 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.923394 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:34Z","lastTransitionTime":"2025-12-09T16:57:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:34 crc kubenswrapper[4954]: I1209 16:57:34.966616 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:57:34 crc kubenswrapper[4954]: E1209 16:57:34.966782 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:58:06.966764206 +0000 UTC m=+83.354938026 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.025335 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.025431 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.025454 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.025487 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.025511 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:35Z","lastTransitionTime":"2025-12-09T16:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.067958 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.068016 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.068042 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.068062 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068130 4954 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068163 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068188 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068202 4954 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068224 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:58:07.068205898 +0000 UTC m=+83.456379708 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068251 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 16:58:07.06823497 +0000 UTC m=+83.456408790 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068148 4954 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068290 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:58:07.068281342 +0000 UTC m=+83.456455162 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068730 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068760 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068774 4954 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.068826 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 16:58:07.068814935 +0000 UTC m=+83.456988805 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.120186 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.120333 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.120412 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.120473 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.120522 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:35 crc kubenswrapper[4954]: E1209 16:57:35.120576 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.128347 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.128379 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.128388 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.128402 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.128414 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:35Z","lastTransitionTime":"2025-12-09T16:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.231734 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.232103 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.232186 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.232289 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.232425 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:35Z","lastTransitionTime":"2025-12-09T16:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.335304 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.335831 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.335913 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.336026 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.336111 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:35Z","lastTransitionTime":"2025-12-09T16:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.438929 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.438987 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.438995 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.439020 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.439031 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:35Z","lastTransitionTime":"2025-12-09T16:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.541937 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.541995 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.542014 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.542041 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.542060 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:35Z","lastTransitionTime":"2025-12-09T16:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.645307 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.645348 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.645357 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.645371 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.645381 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:35Z","lastTransitionTime":"2025-12-09T16:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.748165 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.748204 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.748213 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.748227 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.748238 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:35Z","lastTransitionTime":"2025-12-09T16:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.850576 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.850639 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.850651 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.850670 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.850680 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:35Z","lastTransitionTime":"2025-12-09T16:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.952482 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.952532 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.952543 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.952557 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:35 crc kubenswrapper[4954]: I1209 16:57:35.952570 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:35Z","lastTransitionTime":"2025-12-09T16:57:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.055170 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.055205 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.055213 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.055227 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.055240 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.119471 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:36 crc kubenswrapper[4954]: E1209 16:57:36.119647 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.157691 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.157726 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.157734 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.157749 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.157760 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.260077 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.260117 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.260129 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.260145 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.260155 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.362945 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.363233 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.363329 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.363406 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.363484 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.429431 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.429470 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.429482 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.429504 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.429517 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: E1209 16:57:36.441446 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:36Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.445316 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.445348 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.445357 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.445373 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.445382 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: E1209 16:57:36.457037 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:36Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.460611 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.460787 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.460902 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.460988 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.461073 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: E1209 16:57:36.472605 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:36Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.479152 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.479192 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.479206 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.479222 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.479233 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: E1209 16:57:36.492400 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:36Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.495925 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.495947 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.495955 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.495968 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.495980 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: E1209 16:57:36.509190 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:36Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:36 crc kubenswrapper[4954]: E1209 16:57:36.509390 4954 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.511121 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.511157 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.511169 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.511184 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.511195 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.613790 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.613834 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.613845 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.613862 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.613873 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.716321 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.716611 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.716698 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.716789 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.716867 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.818811 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.818849 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.818860 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.818876 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.818887 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.921800 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.921857 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.921876 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.921903 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:36 crc kubenswrapper[4954]: I1209 16:57:36.921920 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:36Z","lastTransitionTime":"2025-12-09T16:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.024113 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.024425 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.024499 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.024566 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.024684 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:37Z","lastTransitionTime":"2025-12-09T16:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.119263 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.119300 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:37 crc kubenswrapper[4954]: E1209 16:57:37.119511 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:37 crc kubenswrapper[4954]: E1209 16:57:37.119693 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.119298 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:37 crc kubenswrapper[4954]: E1209 16:57:37.120249 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.127538 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.127620 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.127641 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.127666 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.127686 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:37Z","lastTransitionTime":"2025-12-09T16:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.230790 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.230823 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.230835 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.230849 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.230857 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:37Z","lastTransitionTime":"2025-12-09T16:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.333808 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.333876 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.333899 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.333933 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.333955 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:37Z","lastTransitionTime":"2025-12-09T16:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.436570 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.436635 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.436647 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.436662 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.436673 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:37Z","lastTransitionTime":"2025-12-09T16:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.539501 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.539550 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.539562 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.539581 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.539609 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:37Z","lastTransitionTime":"2025-12-09T16:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.641513 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.641555 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.641566 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.641583 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.641618 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:37Z","lastTransitionTime":"2025-12-09T16:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.743213 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.743254 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.743264 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.743279 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.743288 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:37Z","lastTransitionTime":"2025-12-09T16:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.845887 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.845932 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.845943 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.845958 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.845969 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:37Z","lastTransitionTime":"2025-12-09T16:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.949098 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.949153 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.949165 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.949182 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:37 crc kubenswrapper[4954]: I1209 16:57:37.949191 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:37Z","lastTransitionTime":"2025-12-09T16:57:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.052525 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.052572 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.052584 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.052617 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.052629 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:38Z","lastTransitionTime":"2025-12-09T16:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.119924 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:38 crc kubenswrapper[4954]: E1209 16:57:38.120060 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.154905 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.155204 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.155286 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.155476 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.155613 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:38Z","lastTransitionTime":"2025-12-09T16:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.258558 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.258616 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.258628 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.258648 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.258659 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:38Z","lastTransitionTime":"2025-12-09T16:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.360857 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.360905 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.360917 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.360935 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.360944 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:38Z","lastTransitionTime":"2025-12-09T16:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.464057 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.464106 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.464122 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.464140 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.464149 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:38Z","lastTransitionTime":"2025-12-09T16:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.567098 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.567143 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.567154 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.567174 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.567185 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:38Z","lastTransitionTime":"2025-12-09T16:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.669720 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.669758 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.669769 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.669783 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.669791 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:38Z","lastTransitionTime":"2025-12-09T16:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.772348 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.772386 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.772396 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.772413 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.772426 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:38Z","lastTransitionTime":"2025-12-09T16:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.875667 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.875727 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.875739 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.875761 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.875774 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:38Z","lastTransitionTime":"2025-12-09T16:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.978180 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.978236 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.978252 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.978273 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:38 crc kubenswrapper[4954]: I1209 16:57:38.978286 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:38Z","lastTransitionTime":"2025-12-09T16:57:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.081628 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.081743 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.081760 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.081790 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.081807 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:39Z","lastTransitionTime":"2025-12-09T16:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.120091 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.120156 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.120133 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:39 crc kubenswrapper[4954]: E1209 16:57:39.120294 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:39 crc kubenswrapper[4954]: E1209 16:57:39.120426 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:39 crc kubenswrapper[4954]: E1209 16:57:39.120797 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.184840 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.184901 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.184923 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.184958 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.184980 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:39Z","lastTransitionTime":"2025-12-09T16:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.287945 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.287999 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.288013 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.288034 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.288048 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:39Z","lastTransitionTime":"2025-12-09T16:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.390412 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.390447 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.390459 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.390473 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.390482 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:39Z","lastTransitionTime":"2025-12-09T16:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.492757 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.492828 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.492843 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.492881 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.492895 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:39Z","lastTransitionTime":"2025-12-09T16:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.595420 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.595465 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.595480 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.595499 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.595513 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:39Z","lastTransitionTime":"2025-12-09T16:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.699502 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.699579 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.699606 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.699624 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.699637 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:39Z","lastTransitionTime":"2025-12-09T16:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.803150 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.803201 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.803211 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.803229 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.803239 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:39Z","lastTransitionTime":"2025-12-09T16:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.906912 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.906959 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.906971 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.906991 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:39 crc kubenswrapper[4954]: I1209 16:57:39.907004 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:39Z","lastTransitionTime":"2025-12-09T16:57:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.010850 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.010915 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.010927 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.010945 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.010954 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:40Z","lastTransitionTime":"2025-12-09T16:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.114113 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.114164 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.114175 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.114200 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.114213 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:40Z","lastTransitionTime":"2025-12-09T16:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.119840 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:40 crc kubenswrapper[4954]: E1209 16:57:40.119998 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.216953 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.217004 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.217014 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.217034 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.217046 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:40Z","lastTransitionTime":"2025-12-09T16:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.319887 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.319917 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.319925 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.319940 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.319950 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:40Z","lastTransitionTime":"2025-12-09T16:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.422763 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.422807 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.422817 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.422835 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.422846 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:40Z","lastTransitionTime":"2025-12-09T16:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.526105 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.526176 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.526196 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.526225 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.526246 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:40Z","lastTransitionTime":"2025-12-09T16:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.630253 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.630366 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.630400 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.630435 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.630457 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:40Z","lastTransitionTime":"2025-12-09T16:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.734169 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.734256 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.734277 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.734311 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.734333 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:40Z","lastTransitionTime":"2025-12-09T16:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.838418 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.838469 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.838482 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.838503 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.838517 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:40Z","lastTransitionTime":"2025-12-09T16:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.942564 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.942677 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.942699 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.942736 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:40 crc kubenswrapper[4954]: I1209 16:57:40.942760 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:40Z","lastTransitionTime":"2025-12-09T16:57:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.046412 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.046485 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.046504 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.046535 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.046557 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:41Z","lastTransitionTime":"2025-12-09T16:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.119861 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.119899 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.119996 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:41 crc kubenswrapper[4954]: E1209 16:57:41.120071 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:41 crc kubenswrapper[4954]: E1209 16:57:41.120261 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:41 crc kubenswrapper[4954]: E1209 16:57:41.120405 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.157295 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.157375 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.157394 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.157428 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.157448 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:41Z","lastTransitionTime":"2025-12-09T16:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.260660 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.261864 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.262022 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.262199 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.262352 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:41Z","lastTransitionTime":"2025-12-09T16:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.365918 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.366015 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.366031 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.366060 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.366082 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:41Z","lastTransitionTime":"2025-12-09T16:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.469205 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.469269 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.469283 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.469305 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.469320 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:41Z","lastTransitionTime":"2025-12-09T16:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.571620 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.571680 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.571693 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.571717 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.571732 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:41Z","lastTransitionTime":"2025-12-09T16:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.675866 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.675931 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.675944 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.675968 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.675981 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:41Z","lastTransitionTime":"2025-12-09T16:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.778895 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.778955 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.778967 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.778990 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.779007 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:41Z","lastTransitionTime":"2025-12-09T16:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.881486 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.881550 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.881561 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.881574 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.881584 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:41Z","lastTransitionTime":"2025-12-09T16:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.984526 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.984613 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.984633 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.984660 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:41 crc kubenswrapper[4954]: I1209 16:57:41.984678 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:41Z","lastTransitionTime":"2025-12-09T16:57:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.088328 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.088377 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.088390 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.088410 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.088422 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:42Z","lastTransitionTime":"2025-12-09T16:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.119533 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:42 crc kubenswrapper[4954]: E1209 16:57:42.119764 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.122085 4954 scope.go:117] "RemoveContainer" containerID="c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7" Dec 09 16:57:42 crc kubenswrapper[4954]: E1209 16:57:42.122541 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.192209 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.192281 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.192297 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.192323 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.192376 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:42Z","lastTransitionTime":"2025-12-09T16:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.296577 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.296703 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.296725 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.296764 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.296789 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:42Z","lastTransitionTime":"2025-12-09T16:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.400836 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.400895 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.400911 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.400934 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.400952 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:42Z","lastTransitionTime":"2025-12-09T16:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.504204 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.504277 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.504299 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.504331 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.504349 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:42Z","lastTransitionTime":"2025-12-09T16:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.608137 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.608188 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.608199 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.608221 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.608234 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:42Z","lastTransitionTime":"2025-12-09T16:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.710427 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.710699 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.710768 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.710830 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.710890 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:42Z","lastTransitionTime":"2025-12-09T16:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.813903 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.813948 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.814006 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.814029 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.814041 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:42Z","lastTransitionTime":"2025-12-09T16:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.917231 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.917704 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.917815 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.917899 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:42 crc kubenswrapper[4954]: I1209 16:57:42.917971 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:42Z","lastTransitionTime":"2025-12-09T16:57:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.020951 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.021298 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.021643 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.021969 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.022222 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:43Z","lastTransitionTime":"2025-12-09T16:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.120254 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.120413 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.120461 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:43 crc kubenswrapper[4954]: E1209 16:57:43.121302 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:43 crc kubenswrapper[4954]: E1209 16:57:43.121430 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:43 crc kubenswrapper[4954]: E1209 16:57:43.121574 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.125757 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.125831 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.125853 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.125913 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.125933 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:43Z","lastTransitionTime":"2025-12-09T16:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.228981 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.229028 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.229038 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.229058 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.229068 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:43Z","lastTransitionTime":"2025-12-09T16:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.333330 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.334157 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.334256 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.334349 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.334475 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:43Z","lastTransitionTime":"2025-12-09T16:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.438448 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.439083 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.439291 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.439465 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.439626 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:43Z","lastTransitionTime":"2025-12-09T16:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.543243 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.543297 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.543311 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.543335 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.543349 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:43Z","lastTransitionTime":"2025-12-09T16:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.645910 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.645943 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.645952 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.645966 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.645975 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:43Z","lastTransitionTime":"2025-12-09T16:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.748808 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.749063 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.749160 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.749265 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.749359 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:43Z","lastTransitionTime":"2025-12-09T16:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.852689 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.852756 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.852770 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.852790 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.852803 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:43Z","lastTransitionTime":"2025-12-09T16:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.956500 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.956560 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.956574 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.956614 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:43 crc kubenswrapper[4954]: I1209 16:57:43.956631 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:43Z","lastTransitionTime":"2025-12-09T16:57:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.060350 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.060722 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.060790 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.060873 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.060937 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:44Z","lastTransitionTime":"2025-12-09T16:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.119224 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:44 crc kubenswrapper[4954]: E1209 16:57:44.120126 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.138689 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e690f64-775b-4059-9819-18cc3c6da1c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327ec4b9b835a025b2efad8b30d5eab591d0860da7b3a214553b94b9b026c7f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e462de97284638e1161f14f470c25bad53981d0d08f81dde5980991e189edecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23844430c0b5f37c59d048d297bf804c4fbc96cddc4958465f909b24e2635e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.155731 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.163607 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.163644 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.163657 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.163676 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.163689 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:44Z","lastTransitionTime":"2025-12-09T16:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.169725 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.183499 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.196579 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.209797 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.223244 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.239358 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.257312 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.266576 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.266637 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.266649 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.266667 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.266682 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:44Z","lastTransitionTime":"2025-12-09T16:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.275291 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.287679 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.297626 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.315942 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.330369 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.341890 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.355089 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.370133 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.370200 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.370215 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.370233 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.370244 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:44Z","lastTransitionTime":"2025-12-09T16:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.375288 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:28Z\\\",\\\"message\\\":\\\"ce_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:27.963752 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1209 16:57:27.963753 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963755 6583 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-ffkzn in node crc\\\\nI1209 16:57:27.963761 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963765 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963771 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963776 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.388247 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:44Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.473781 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.473899 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.473926 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.473967 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.473997 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:44Z","lastTransitionTime":"2025-12-09T16:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.577568 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.577680 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.577694 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.577718 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.577733 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:44Z","lastTransitionTime":"2025-12-09T16:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.681875 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.681942 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.681966 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.682000 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.682024 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:44Z","lastTransitionTime":"2025-12-09T16:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.786048 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.786502 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.786627 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.786759 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.786868 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:44Z","lastTransitionTime":"2025-12-09T16:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.890455 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.890524 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.890542 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.890573 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.890610 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:44Z","lastTransitionTime":"2025-12-09T16:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.992873 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.992910 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.992918 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.992932 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:44 crc kubenswrapper[4954]: I1209 16:57:44.992941 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:44Z","lastTransitionTime":"2025-12-09T16:57:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.095902 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.095940 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.095951 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.095965 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.095975 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:45Z","lastTransitionTime":"2025-12-09T16:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.119413 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:45 crc kubenswrapper[4954]: E1209 16:57:45.119581 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.119799 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.119848 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:45 crc kubenswrapper[4954]: E1209 16:57:45.119939 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:45 crc kubenswrapper[4954]: E1209 16:57:45.119999 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.198150 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.198188 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.198201 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.198217 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.198229 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:45Z","lastTransitionTime":"2025-12-09T16:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.301628 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.301679 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.301691 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.301711 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.301723 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:45Z","lastTransitionTime":"2025-12-09T16:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.403912 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.403955 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.403964 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.403978 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.403987 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:45Z","lastTransitionTime":"2025-12-09T16:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.506374 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.506425 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.506434 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.506448 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.506456 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:45Z","lastTransitionTime":"2025-12-09T16:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.609137 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.609182 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.609191 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.609207 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.609218 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:45Z","lastTransitionTime":"2025-12-09T16:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.711843 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.711891 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.711905 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.711922 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.711950 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:45Z","lastTransitionTime":"2025-12-09T16:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.813985 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.814024 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.814035 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.814058 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.814069 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:45Z","lastTransitionTime":"2025-12-09T16:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.917434 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.917489 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.917504 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.917527 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:45 crc kubenswrapper[4954]: I1209 16:57:45.917543 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:45Z","lastTransitionTime":"2025-12-09T16:57:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.020707 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.020744 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.020753 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.020767 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.020776 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.119968 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:46 crc kubenswrapper[4954]: E1209 16:57:46.120124 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.123301 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.123357 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.123376 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.123441 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.123452 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.226227 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.226262 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.226271 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.226284 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.226294 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.328266 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.328309 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.328318 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.328333 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.328342 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.430287 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.430317 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.430326 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.430342 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.430350 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.532123 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.532159 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.532169 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.532185 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.532196 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.618536 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.618563 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.618571 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.618583 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.618608 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: E1209 16:57:46.629026 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:46Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.632182 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.632207 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.632216 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.632230 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.632241 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: E1209 16:57:46.642725 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:46Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.645333 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.645356 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.645365 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.645378 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.645386 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: E1209 16:57:46.655534 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:46Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.659331 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.659368 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.659382 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.659398 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.659409 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: E1209 16:57:46.671955 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:46Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.675983 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.676078 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.676109 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.676158 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.676195 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: E1209 16:57:46.699208 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:46Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:46 crc kubenswrapper[4954]: E1209 16:57:46.699513 4954 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.702097 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.702138 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.702150 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.702198 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.702216 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.805345 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.805383 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.805393 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.805412 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.805426 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.909022 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.909078 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.909087 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.909109 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:46 crc kubenswrapper[4954]: I1209 16:57:46.909121 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:46Z","lastTransitionTime":"2025-12-09T16:57:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.011639 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.011676 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.011687 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.011710 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.011726 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:47Z","lastTransitionTime":"2025-12-09T16:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.114823 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.114865 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.114874 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.114889 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.114899 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:47Z","lastTransitionTime":"2025-12-09T16:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.119254 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.119434 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.119495 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:47 crc kubenswrapper[4954]: E1209 16:57:47.119445 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:47 crc kubenswrapper[4954]: E1209 16:57:47.119626 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:47 crc kubenswrapper[4954]: E1209 16:57:47.119831 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.218276 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.218349 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.218566 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.218626 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.218646 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:47Z","lastTransitionTime":"2025-12-09T16:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.321246 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.321295 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.321304 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.321318 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.321328 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:47Z","lastTransitionTime":"2025-12-09T16:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.423525 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.423562 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.423571 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.423586 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.423624 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:47Z","lastTransitionTime":"2025-12-09T16:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.525832 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.525874 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.525887 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.525902 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.525912 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:47Z","lastTransitionTime":"2025-12-09T16:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.629189 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.629246 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.629258 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.629279 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.629291 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:47Z","lastTransitionTime":"2025-12-09T16:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.732063 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.732104 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.732115 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.732136 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.732147 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:47Z","lastTransitionTime":"2025-12-09T16:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.834272 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.834364 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.834380 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.834410 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.834427 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:47Z","lastTransitionTime":"2025-12-09T16:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.938293 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.938339 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.938347 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.938365 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:47 crc kubenswrapper[4954]: I1209 16:57:47.938375 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:47Z","lastTransitionTime":"2025-12-09T16:57:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.040455 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.040515 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.040529 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.040549 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.040563 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:48Z","lastTransitionTime":"2025-12-09T16:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.119963 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:48 crc kubenswrapper[4954]: E1209 16:57:48.120111 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.142848 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.142891 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.142901 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.142916 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.142927 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:48Z","lastTransitionTime":"2025-12-09T16:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.245958 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.245996 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.246006 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.246024 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.246034 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:48Z","lastTransitionTime":"2025-12-09T16:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.348933 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.348979 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.348989 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.349007 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.349018 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:48Z","lastTransitionTime":"2025-12-09T16:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.451539 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.451608 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.451619 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.451638 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.451651 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:48Z","lastTransitionTime":"2025-12-09T16:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.519041 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:48 crc kubenswrapper[4954]: E1209 16:57:48.519246 4954 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:48 crc kubenswrapper[4954]: E1209 16:57:48.519325 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs podName:856fc314-24bd-403e-a4b0-3dcd73eba595 nodeName:}" failed. No retries permitted until 2025-12-09 16:58:20.519303531 +0000 UTC m=+96.907477361 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs") pod "network-metrics-daemon-ffkzn" (UID: "856fc314-24bd-403e-a4b0-3dcd73eba595") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.553799 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.553839 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.553850 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.553864 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.553874 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:48Z","lastTransitionTime":"2025-12-09T16:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.656223 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.656261 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.656270 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.656284 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.656295 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:48Z","lastTransitionTime":"2025-12-09T16:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.758829 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.758873 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.758885 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.758901 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.758911 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:48Z","lastTransitionTime":"2025-12-09T16:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.861549 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.861576 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.861587 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.861624 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.861636 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:48Z","lastTransitionTime":"2025-12-09T16:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.964163 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.964200 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.964208 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.964222 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:48 crc kubenswrapper[4954]: I1209 16:57:48.964231 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:48Z","lastTransitionTime":"2025-12-09T16:57:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.067919 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.067987 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.068006 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.068043 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.068068 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:49Z","lastTransitionTime":"2025-12-09T16:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.119792 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.119869 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.119795 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:49 crc kubenswrapper[4954]: E1209 16:57:49.120039 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:49 crc kubenswrapper[4954]: E1209 16:57:49.120176 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:49 crc kubenswrapper[4954]: E1209 16:57:49.120341 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.170703 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.170747 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.170768 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.170793 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.170805 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:49Z","lastTransitionTime":"2025-12-09T16:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.273898 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.273955 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.273973 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.273999 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.274016 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:49Z","lastTransitionTime":"2025-12-09T16:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.377252 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.377308 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.377321 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.377342 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.377355 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:49Z","lastTransitionTime":"2025-12-09T16:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.479755 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.479796 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.479805 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.479823 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.479836 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:49Z","lastTransitionTime":"2025-12-09T16:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.582581 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.582661 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.582673 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.582691 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.582702 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:49Z","lastTransitionTime":"2025-12-09T16:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.685938 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.685999 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.686017 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.686040 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.686053 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:49Z","lastTransitionTime":"2025-12-09T16:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.788347 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.788384 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.788393 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.788408 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.788419 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:49Z","lastTransitionTime":"2025-12-09T16:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.891168 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.891251 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.891265 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.891294 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.891310 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:49Z","lastTransitionTime":"2025-12-09T16:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.994875 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.994934 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.994948 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.994975 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:49 crc kubenswrapper[4954]: I1209 16:57:49.994995 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:49Z","lastTransitionTime":"2025-12-09T16:57:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.097654 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.097703 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.097713 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.097729 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.097739 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:50Z","lastTransitionTime":"2025-12-09T16:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.119254 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:50 crc kubenswrapper[4954]: E1209 16:57:50.119470 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.200305 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.200340 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.200350 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.200366 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.200386 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:50Z","lastTransitionTime":"2025-12-09T16:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.305186 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.305267 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.305470 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.305490 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.305862 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:50Z","lastTransitionTime":"2025-12-09T16:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.409047 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.409101 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.409112 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.409132 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.409143 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:50Z","lastTransitionTime":"2025-12-09T16:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.511665 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.511717 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.511727 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.511747 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.511760 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:50Z","lastTransitionTime":"2025-12-09T16:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.533340 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9n2h_1eccf2e0-30a3-4201-a23f-85a92a522d72/kube-multus/0.log" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.533420 4954 generic.go:334] "Generic (PLEG): container finished" podID="1eccf2e0-30a3-4201-a23f-85a92a522d72" containerID="3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c" exitCode=1 Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.533468 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9n2h" event={"ID":"1eccf2e0-30a3-4201-a23f-85a92a522d72","Type":"ContainerDied","Data":"3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c"} Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.534010 4954 scope.go:117] "RemoveContainer" containerID="3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.550816 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.566117 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.580239 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.593160 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.613802 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.613990 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.614114 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.614220 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.614315 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:50Z","lastTransitionTime":"2025-12-09T16:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.614964 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.629721 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.641541 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.654511 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.675118 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:28Z\\\",\\\"message\\\":\\\"ce_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:27.963752 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1209 16:57:27.963753 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963755 6583 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-ffkzn in node crc\\\\nI1209 16:57:27.963761 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963765 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963771 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963776 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.686264 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.699274 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e690f64-775b-4059-9819-18cc3c6da1c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327ec4b9b835a025b2efad8b30d5eab591d0860da7b3a214553b94b9b026c7f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e462de97284638e1161f14f470c25bad53981d0d08f81dde5980991e189edecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23844430c0b5f37c59d048d297bf804c4fbc96cddc4958465f909b24e2635e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.715468 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:49Z\\\",\\\"message\\\":\\\"2025-12-09T16:57:04+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5\\\\n2025-12-09T16:57:04+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5 to /host/opt/cni/bin/\\\\n2025-12-09T16:57:04Z [verbose] multus-daemon started\\\\n2025-12-09T16:57:04Z [verbose] Readiness Indicator file check\\\\n2025-12-09T16:57:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.716800 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.716829 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.716838 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.716853 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.716862 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:50Z","lastTransitionTime":"2025-12-09T16:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.731489 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.743763 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.758882 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.772436 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.785555 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.800900 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:50Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.819941 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.819987 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.820001 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.820022 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.820036 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:50Z","lastTransitionTime":"2025-12-09T16:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.922630 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.922687 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.922701 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.922725 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:50 crc kubenswrapper[4954]: I1209 16:57:50.922739 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:50Z","lastTransitionTime":"2025-12-09T16:57:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.025632 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.025678 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.025688 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.025706 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.025719 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:51Z","lastTransitionTime":"2025-12-09T16:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.120134 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.120163 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:51 crc kubenswrapper[4954]: E1209 16:57:51.120360 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.120191 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:51 crc kubenswrapper[4954]: E1209 16:57:51.120427 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:51 crc kubenswrapper[4954]: E1209 16:57:51.120489 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.128819 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.128857 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.128873 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.128894 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.128907 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:51Z","lastTransitionTime":"2025-12-09T16:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.231472 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.231529 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.231542 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.231556 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.231566 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:51Z","lastTransitionTime":"2025-12-09T16:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.333964 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.334002 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.334012 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.334024 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.334033 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:51Z","lastTransitionTime":"2025-12-09T16:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.436627 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.436667 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.436677 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.436698 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.436709 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:51Z","lastTransitionTime":"2025-12-09T16:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.537456 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9n2h_1eccf2e0-30a3-4201-a23f-85a92a522d72/kube-multus/0.log" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.537527 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9n2h" event={"ID":"1eccf2e0-30a3-4201-a23f-85a92a522d72","Type":"ContainerStarted","Data":"adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.537980 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.538017 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.538029 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.538045 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.538054 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:51Z","lastTransitionTime":"2025-12-09T16:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.549727 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.566908 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:28Z\\\",\\\"message\\\":\\\"ce_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:27.963752 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1209 16:57:27.963753 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963755 6583 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-ffkzn in node crc\\\\nI1209 16:57:27.963761 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963765 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963771 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963776 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.576943 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.593294 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.606037 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.615363 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.624649 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.633853 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.640343 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.640547 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.640677 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.640766 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.640850 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:51Z","lastTransitionTime":"2025-12-09T16:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.644035 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.653861 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e690f64-775b-4059-9819-18cc3c6da1c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327ec4b9b835a025b2efad8b30d5eab591d0860da7b3a214553b94b9b026c7f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e462de97284638e1161f14f470c25bad53981d0d08f81dde5980991e189edecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23844430c0b5f37c59d048d297bf804c4fbc96cddc4958465f909b24e2635e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.665857 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:49Z\\\",\\\"message\\\":\\\"2025-12-09T16:57:04+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5\\\\n2025-12-09T16:57:04+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5 to /host/opt/cni/bin/\\\\n2025-12-09T16:57:04Z [verbose] multus-daemon started\\\\n2025-12-09T16:57:04Z [verbose] Readiness Indicator file check\\\\n2025-12-09T16:57:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.679685 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.693768 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.705290 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.718266 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.730876 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.743464 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.743700 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.743723 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.743732 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.743746 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.743756 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:51Z","lastTransitionTime":"2025-12-09T16:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.755000 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:51Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.846563 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.846651 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.846668 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.846690 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.846706 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:51Z","lastTransitionTime":"2025-12-09T16:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.948346 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.948659 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.948807 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.948907 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:51 crc kubenswrapper[4954]: I1209 16:57:51.948997 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:51Z","lastTransitionTime":"2025-12-09T16:57:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.052224 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.052262 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.052273 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.052287 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.052297 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:52Z","lastTransitionTime":"2025-12-09T16:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.119550 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:52 crc kubenswrapper[4954]: E1209 16:57:52.119786 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.134987 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.154492 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.154521 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.154529 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.154541 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.154549 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:52Z","lastTransitionTime":"2025-12-09T16:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.256804 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.256834 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.256842 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.256856 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.256865 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:52Z","lastTransitionTime":"2025-12-09T16:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.358924 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.358950 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.358960 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.358973 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.358981 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:52Z","lastTransitionTime":"2025-12-09T16:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.460918 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.460950 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.460958 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.460972 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.460981 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:52Z","lastTransitionTime":"2025-12-09T16:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.564642 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.564694 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.564708 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.564731 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.564745 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:52Z","lastTransitionTime":"2025-12-09T16:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.668128 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.668210 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.668229 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.668268 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.668293 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:52Z","lastTransitionTime":"2025-12-09T16:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.771641 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.772108 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.772123 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.772142 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.772160 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:52Z","lastTransitionTime":"2025-12-09T16:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.874813 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.874853 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.874861 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.874875 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.874885 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:52Z","lastTransitionTime":"2025-12-09T16:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.977184 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.977231 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.977240 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.977255 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:52 crc kubenswrapper[4954]: I1209 16:57:52.977268 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:52Z","lastTransitionTime":"2025-12-09T16:57:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.080076 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.080134 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.080147 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.080163 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.080174 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:53Z","lastTransitionTime":"2025-12-09T16:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.119885 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.119934 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.119945 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:53 crc kubenswrapper[4954]: E1209 16:57:53.120029 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:53 crc kubenswrapper[4954]: E1209 16:57:53.120096 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:53 crc kubenswrapper[4954]: E1209 16:57:53.120174 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.182969 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.183007 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.183046 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.183063 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.183072 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:53Z","lastTransitionTime":"2025-12-09T16:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.284872 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.284909 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.284919 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.284934 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.284945 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:53Z","lastTransitionTime":"2025-12-09T16:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.386990 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.387042 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.387053 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.387070 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.387082 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:53Z","lastTransitionTime":"2025-12-09T16:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.489953 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.490026 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.490040 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.490061 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.490073 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:53Z","lastTransitionTime":"2025-12-09T16:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.592418 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.592457 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.592614 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.592630 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.592644 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:53Z","lastTransitionTime":"2025-12-09T16:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.694940 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.695530 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.695652 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.695734 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.695792 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:53Z","lastTransitionTime":"2025-12-09T16:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.798446 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.798496 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.798509 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.798523 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.798532 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:53Z","lastTransitionTime":"2025-12-09T16:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.901671 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.901709 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.901721 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.901740 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:53 crc kubenswrapper[4954]: I1209 16:57:53.901751 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:53Z","lastTransitionTime":"2025-12-09T16:57:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.003901 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.004184 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.004268 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.004362 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.004450 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:54Z","lastTransitionTime":"2025-12-09T16:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.106631 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.106661 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.106669 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.106682 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.106692 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:54Z","lastTransitionTime":"2025-12-09T16:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.120152 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:54 crc kubenswrapper[4954]: E1209 16:57:54.120267 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.132881 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.141879 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.151766 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.162436 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d5afb1-2789-4058-97bf-b4604a9a5234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17199e1bc8be0a9cb3e83dca2669f2d58187f64c1b82d3bf4a84b1a39538ec03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.177848 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e690f64-775b-4059-9819-18cc3c6da1c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327ec4b9b835a025b2efad8b30d5eab591d0860da7b3a214553b94b9b026c7f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e462de97284638e1161f14f470c25bad53981d0d08f81dde5980991e189edecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23844430c0b5f37c59d048d297bf804c4fbc96cddc4958465f909b24e2635e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.191108 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:49Z\\\",\\\"message\\\":\\\"2025-12-09T16:57:04+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5\\\\n2025-12-09T16:57:04+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5 to /host/opt/cni/bin/\\\\n2025-12-09T16:57:04Z [verbose] multus-daemon started\\\\n2025-12-09T16:57:04Z [verbose] Readiness Indicator file check\\\\n2025-12-09T16:57:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.204883 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.208889 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.208928 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.208939 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.208953 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.208966 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:54Z","lastTransitionTime":"2025-12-09T16:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.219251 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.237094 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.254774 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.271468 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.284046 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.294247 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.307808 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.311924 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.312087 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.312169 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.312252 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.312330 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:54Z","lastTransitionTime":"2025-12-09T16:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.327106 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:28Z\\\",\\\"message\\\":\\\"ce_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:27.963752 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1209 16:57:27.963753 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963755 6583 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-ffkzn in node crc\\\\nI1209 16:57:27.963761 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963765 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963771 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963776 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.337074 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.354728 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.367671 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.378502 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:54Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.415358 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.415410 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.415424 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.415442 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.415454 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:54Z","lastTransitionTime":"2025-12-09T16:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.517659 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.517728 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.517749 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.517778 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.517802 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:54Z","lastTransitionTime":"2025-12-09T16:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.620442 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.620748 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.620850 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.620936 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.621013 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:54Z","lastTransitionTime":"2025-12-09T16:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.723513 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.723548 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.723556 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.723571 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.723583 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:54Z","lastTransitionTime":"2025-12-09T16:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.825812 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.825859 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.825871 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.825889 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.825901 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:54Z","lastTransitionTime":"2025-12-09T16:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.928413 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.928458 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.928467 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.928482 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:54 crc kubenswrapper[4954]: I1209 16:57:54.928490 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:54Z","lastTransitionTime":"2025-12-09T16:57:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.031578 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.031882 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.032028 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.032122 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.032238 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:55Z","lastTransitionTime":"2025-12-09T16:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.119977 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:55 crc kubenswrapper[4954]: E1209 16:57:55.120488 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.120016 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:55 crc kubenswrapper[4954]: E1209 16:57:55.121153 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.119985 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:55 crc kubenswrapper[4954]: E1209 16:57:55.121784 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.135401 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.135868 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.135979 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.136049 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.136126 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:55Z","lastTransitionTime":"2025-12-09T16:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.238672 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.238735 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.238746 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.238767 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.238781 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:55Z","lastTransitionTime":"2025-12-09T16:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.349641 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.349675 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.349685 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.349699 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.349711 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:55Z","lastTransitionTime":"2025-12-09T16:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.452634 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.452702 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.452717 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.452746 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.452760 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:55Z","lastTransitionTime":"2025-12-09T16:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.554801 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.554839 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.554847 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.554862 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.554871 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:55Z","lastTransitionTime":"2025-12-09T16:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.656963 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.656996 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.657006 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.657021 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.657035 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:55Z","lastTransitionTime":"2025-12-09T16:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.760002 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.760041 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.760049 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.760068 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.760076 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:55Z","lastTransitionTime":"2025-12-09T16:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.862207 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.862248 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.862267 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.862283 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.862293 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:55Z","lastTransitionTime":"2025-12-09T16:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.964215 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.964262 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.964274 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.964293 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:55 crc kubenswrapper[4954]: I1209 16:57:55.964305 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:55Z","lastTransitionTime":"2025-12-09T16:57:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.067058 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.067103 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.067114 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.067129 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.067138 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.120188 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:56 crc kubenswrapper[4954]: E1209 16:57:56.120388 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.169154 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.169190 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.169201 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.169215 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.169226 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.272010 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.272038 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.272045 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.272059 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.272067 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.375119 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.375158 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.375170 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.375192 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.375203 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.477645 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.477677 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.477686 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.477700 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.477710 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.580456 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.580496 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.580506 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.580523 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.580535 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.682863 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.682947 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.682960 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.682977 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.682989 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.785234 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.785272 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.785284 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.785301 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.785312 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.887909 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.887979 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.887991 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.888020 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.888037 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.992776 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.993070 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.993141 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.993230 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.993302 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.999729 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.999831 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.999915 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:56 crc kubenswrapper[4954]: I1209 16:57:56.999983 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.000038 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:56Z","lastTransitionTime":"2025-12-09T16:57:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: E1209 16:57:57.011443 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:57Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.017684 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.017727 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.017742 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.017766 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.017783 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: E1209 16:57:57.032857 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:57Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.037756 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.038004 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.038096 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.038223 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.038304 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: E1209 16:57:57.051608 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:57Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.056305 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.056351 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.056361 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.056380 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.056390 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: E1209 16:57:57.066780 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:57Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.069747 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.069780 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.069794 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.069811 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.069823 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: E1209 16:57:57.080137 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:57Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:57 crc kubenswrapper[4954]: E1209 16:57:57.080247 4954 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.095881 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.095914 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.095923 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.095940 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.095949 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.119472 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.119514 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:57 crc kubenswrapper[4954]: E1209 16:57:57.119613 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.119699 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:57 crc kubenswrapper[4954]: E1209 16:57:57.119702 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:57 crc kubenswrapper[4954]: E1209 16:57:57.120065 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.120442 4954 scope.go:117] "RemoveContainer" containerID="c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.198322 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.198361 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.198372 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.198387 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.198412 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.301468 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.301520 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.301537 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.301560 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.301577 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.404019 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.404056 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.404065 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.404077 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.404086 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.507157 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.507209 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.507253 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.507275 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.507287 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.609164 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.609200 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.609214 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.609232 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.609242 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.711640 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.711677 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.711686 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.711701 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.711712 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.814373 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.814439 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.814462 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.814493 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.814514 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.916561 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.916622 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.916635 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.916652 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:57 crc kubenswrapper[4954]: I1209 16:57:57.916664 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:57Z","lastTransitionTime":"2025-12-09T16:57:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.019294 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.019337 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.019351 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.019370 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.019382 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:58Z","lastTransitionTime":"2025-12-09T16:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.119671 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:57:58 crc kubenswrapper[4954]: E1209 16:57:58.119790 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.120812 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.120842 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.120850 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.120868 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.120876 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:58Z","lastTransitionTime":"2025-12-09T16:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.223131 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.223168 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.223177 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.223192 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.223202 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:58Z","lastTransitionTime":"2025-12-09T16:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.325442 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.325480 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.325492 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.325508 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.325519 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:58Z","lastTransitionTime":"2025-12-09T16:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.427952 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.427992 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.428003 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.428020 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.428031 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:58Z","lastTransitionTime":"2025-12-09T16:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.530027 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.530071 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.530081 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.530098 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.530109 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:58Z","lastTransitionTime":"2025-12-09T16:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.560535 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/2.log" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.562887 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.563347 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.574246 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.585330 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.595893 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d5afb1-2789-4058-97bf-b4604a9a5234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17199e1bc8be0a9cb3e83dca2669f2d58187f64c1b82d3bf4a84b1a39538ec03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.607192 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e690f64-775b-4059-9819-18cc3c6da1c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327ec4b9b835a025b2efad8b30d5eab591d0860da7b3a214553b94b9b026c7f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e462de97284638e1161f14f470c25bad53981d0d08f81dde5980991e189edecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23844430c0b5f37c59d048d297bf804c4fbc96cddc4958465f909b24e2635e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.621068 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:49Z\\\",\\\"message\\\":\\\"2025-12-09T16:57:04+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5\\\\n2025-12-09T16:57:04+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5 to /host/opt/cni/bin/\\\\n2025-12-09T16:57:04Z [verbose] multus-daemon started\\\\n2025-12-09T16:57:04Z [verbose] Readiness Indicator file check\\\\n2025-12-09T16:57:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.632616 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.632651 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.632660 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.632673 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.632682 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:58Z","lastTransitionTime":"2025-12-09T16:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.633715 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.650811 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.663998 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.685424 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.708413 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.732624 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.735182 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.735217 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.735227 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.735240 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.735249 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:58Z","lastTransitionTime":"2025-12-09T16:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.748683 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.760730 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.779847 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:28Z\\\",\\\"message\\\":\\\"ce_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:27.963752 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1209 16:57:27.963753 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963755 6583 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-ffkzn in node crc\\\\nI1209 16:57:27.963761 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963765 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963771 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963776 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.790629 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.811019 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.825936 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.837873 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.837932 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.837945 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.837966 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.837980 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:58Z","lastTransitionTime":"2025-12-09T16:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.840242 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.852875 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:58Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.940440 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.940487 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.940497 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.940513 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:58 crc kubenswrapper[4954]: I1209 16:57:58.940522 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:58Z","lastTransitionTime":"2025-12-09T16:57:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.042609 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.042686 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.042697 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.042721 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.042733 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:59Z","lastTransitionTime":"2025-12-09T16:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.120027 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.120068 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.120042 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:57:59 crc kubenswrapper[4954]: E1209 16:57:59.120157 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:57:59 crc kubenswrapper[4954]: E1209 16:57:59.120229 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:57:59 crc kubenswrapper[4954]: E1209 16:57:59.120286 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.145351 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.145387 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.145396 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.145412 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.145423 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:59Z","lastTransitionTime":"2025-12-09T16:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.248016 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.248060 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.248072 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.248091 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.248104 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:59Z","lastTransitionTime":"2025-12-09T16:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.354075 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.354121 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.354133 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.354150 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.354160 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:59Z","lastTransitionTime":"2025-12-09T16:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.456863 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.456913 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.456933 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.456960 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.456982 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:59Z","lastTransitionTime":"2025-12-09T16:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.559439 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.559470 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.559478 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.559491 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.559499 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:59Z","lastTransitionTime":"2025-12-09T16:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.566310 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/3.log" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.566795 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/2.log" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.568690 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3" exitCode=1 Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.568718 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.568745 4954 scope.go:117] "RemoveContainer" containerID="c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.569350 4954 scope.go:117] "RemoveContainer" containerID="6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3" Dec 09 16:57:59 crc kubenswrapper[4954]: E1209 16:57:59.569476 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.581957 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d5afb1-2789-4058-97bf-b4604a9a5234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17199e1bc8be0a9cb3e83dca2669f2d58187f64c1b82d3bf4a84b1a39538ec03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.594204 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e690f64-775b-4059-9819-18cc3c6da1c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327ec4b9b835a025b2efad8b30d5eab591d0860da7b3a214553b94b9b026c7f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e462de97284638e1161f14f470c25bad53981d0d08f81dde5980991e189edecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23844430c0b5f37c59d048d297bf804c4fbc96cddc4958465f909b24e2635e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.607864 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:49Z\\\",\\\"message\\\":\\\"2025-12-09T16:57:04+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5\\\\n2025-12-09T16:57:04+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5 to /host/opt/cni/bin/\\\\n2025-12-09T16:57:04Z [verbose] multus-daemon started\\\\n2025-12-09T16:57:04Z [verbose] Readiness Indicator file check\\\\n2025-12-09T16:57:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.619556 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.630823 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.640766 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.652870 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.665481 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.665563 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.665580 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.665636 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.665661 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:59Z","lastTransitionTime":"2025-12-09T16:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.768679 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.768715 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.768726 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.768743 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.768754 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:59Z","lastTransitionTime":"2025-12-09T16:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.800328 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.813737 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.827262 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.842196 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.855213 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.864728 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.870805 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.870836 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.870848 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.870864 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.870876 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:59Z","lastTransitionTime":"2025-12-09T16:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.882330 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.894000 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.905325 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.915988 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.933212 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c99e2628a1c3e7d74b5f5479b93e070a87b5aa1d7f508eda0673970e1c3a13f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:28Z\\\",\\\"message\\\":\\\"ce_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 16:57:27.963752 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1209 16:57:27.963753 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963755 6583 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-ffkzn in node crc\\\\nI1209 16:57:27.963761 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-x2fvb\\\\nI1209 16:57:27.963765 6583 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963771 6583 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1209 16:57:27.963776 6583 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:58Z\\\",\\\"message\\\":\\\"n.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.88\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 16:57:58.387939 7006 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc in node crc\\\\nI1209 16:57:58.387945 7006 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc after 0 failed attempt(s)\\\\nI1209 16:57:58.387944 7006 services_controller.go:452] Built service openshift-console-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1209 16:57:58.387952 7006 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1209 16:57:58.387702 7006 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1209 16:57:58.387959 7006 services_controller.go:453] Built service openshift-console-operator/metrics template LB for network=default: []services.LB{}\\\\nI1209 16:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.945089 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:57:59Z is after 2025-08-24T17:21:41Z" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.973329 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.973369 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.973378 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.973393 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:57:59 crc kubenswrapper[4954]: I1209 16:57:59.973403 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:57:59Z","lastTransitionTime":"2025-12-09T16:57:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.076390 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.076447 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.076459 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.076514 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.076530 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:00Z","lastTransitionTime":"2025-12-09T16:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.119131 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:00 crc kubenswrapper[4954]: E1209 16:58:00.119263 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.178546 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.178613 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.178639 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.178655 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.178667 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:00Z","lastTransitionTime":"2025-12-09T16:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.281379 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.281420 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.281430 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.281445 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.281456 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:00Z","lastTransitionTime":"2025-12-09T16:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.383642 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.383697 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.383710 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.383727 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.383737 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:00Z","lastTransitionTime":"2025-12-09T16:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.486298 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.486346 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.486358 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.486383 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.486397 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:00Z","lastTransitionTime":"2025-12-09T16:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.572909 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/3.log" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.576155 4954 scope.go:117] "RemoveContainer" containerID="6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3" Dec 09 16:58:00 crc kubenswrapper[4954]: E1209 16:58:00.576298 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.587235 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.589834 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.589868 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.589879 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.589898 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.589910 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:00Z","lastTransitionTime":"2025-12-09T16:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.604512 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.621976 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:58Z\\\",\\\"message\\\":\\\"n.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.88\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 16:57:58.387939 7006 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc in node crc\\\\nI1209 16:57:58.387945 7006 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc after 0 failed attempt(s)\\\\nI1209 16:57:58.387944 7006 services_controller.go:452] Built service openshift-console-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1209 16:57:58.387952 7006 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1209 16:57:58.387702 7006 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1209 16:57:58.387959 7006 services_controller.go:453] Built service openshift-console-operator/metrics template LB for network=default: []services.LB{}\\\\nI1209 16:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.632757 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.653065 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.666897 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.679939 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:49Z\\\",\\\"message\\\":\\\"2025-12-09T16:57:04+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5\\\\n2025-12-09T16:57:04+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5 to /host/opt/cni/bin/\\\\n2025-12-09T16:57:04Z [verbose] multus-daemon started\\\\n2025-12-09T16:57:04Z [verbose] Readiness Indicator file check\\\\n2025-12-09T16:57:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.690866 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.691921 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.692120 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.692389 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.692631 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.692791 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:00Z","lastTransitionTime":"2025-12-09T16:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.700652 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.711915 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.721419 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d5afb1-2789-4058-97bf-b4604a9a5234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17199e1bc8be0a9cb3e83dca2669f2d58187f64c1b82d3bf4a84b1a39538ec03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.732084 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e690f64-775b-4059-9819-18cc3c6da1c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327ec4b9b835a025b2efad8b30d5eab591d0860da7b3a214553b94b9b026c7f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e462de97284638e1161f14f470c25bad53981d0d08f81dde5980991e189edecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23844430c0b5f37c59d048d297bf804c4fbc96cddc4958465f909b24e2635e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.743671 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.760215 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.778871 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.791904 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.795092 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.795122 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.795133 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.795149 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.795161 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:00Z","lastTransitionTime":"2025-12-09T16:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.801789 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.816469 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.828987 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:00Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.896884 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.896937 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.896952 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.896974 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.896991 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:00Z","lastTransitionTime":"2025-12-09T16:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.999575 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.999671 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.999681 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.999695 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:00 crc kubenswrapper[4954]: I1209 16:58:00.999706 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:00Z","lastTransitionTime":"2025-12-09T16:58:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.102442 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.102723 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.102745 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.102760 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.102771 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:01Z","lastTransitionTime":"2025-12-09T16:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.119992 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:01 crc kubenswrapper[4954]: E1209 16:58:01.120079 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.120124 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:01 crc kubenswrapper[4954]: E1209 16:58:01.120264 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.120314 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:01 crc kubenswrapper[4954]: E1209 16:58:01.120456 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.204469 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.204513 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.204525 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.204540 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.204551 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:01Z","lastTransitionTime":"2025-12-09T16:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.306694 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.306733 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.306745 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.306760 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.306770 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:01Z","lastTransitionTime":"2025-12-09T16:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.409323 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.409365 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.409374 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.409389 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.409398 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:01Z","lastTransitionTime":"2025-12-09T16:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.511723 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.511764 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.511777 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.511795 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.511810 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:01Z","lastTransitionTime":"2025-12-09T16:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.613847 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.613880 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.613888 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.613901 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.613909 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:01Z","lastTransitionTime":"2025-12-09T16:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.716198 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.716232 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.716243 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.716258 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.716269 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:01Z","lastTransitionTime":"2025-12-09T16:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.818973 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.819035 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.819061 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.819090 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.819111 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:01Z","lastTransitionTime":"2025-12-09T16:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.921713 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.921750 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.921760 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.921776 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:01 crc kubenswrapper[4954]: I1209 16:58:01.921788 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:01Z","lastTransitionTime":"2025-12-09T16:58:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.023347 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.023387 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.023397 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.023413 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.023424 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:02Z","lastTransitionTime":"2025-12-09T16:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.119399 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:02 crc kubenswrapper[4954]: E1209 16:58:02.119545 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.125170 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.125220 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.125234 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.125251 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.125263 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:02Z","lastTransitionTime":"2025-12-09T16:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.227741 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.227779 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.227791 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.227807 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.227821 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:02Z","lastTransitionTime":"2025-12-09T16:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.330408 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.330471 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.330489 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.330514 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.330531 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:02Z","lastTransitionTime":"2025-12-09T16:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.433303 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.433361 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.433373 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.433399 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.433415 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:02Z","lastTransitionTime":"2025-12-09T16:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.536569 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.536635 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.536644 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.536659 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.536675 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:02Z","lastTransitionTime":"2025-12-09T16:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.639204 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.639239 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.639247 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.639261 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.639270 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:02Z","lastTransitionTime":"2025-12-09T16:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.741156 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.741404 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.741472 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.741540 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.741624 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:02Z","lastTransitionTime":"2025-12-09T16:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.843795 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.844088 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.844154 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.844230 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.844313 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:02Z","lastTransitionTime":"2025-12-09T16:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.947160 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.947204 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.947215 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.947231 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:02 crc kubenswrapper[4954]: I1209 16:58:02.947242 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:02Z","lastTransitionTime":"2025-12-09T16:58:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.049511 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.049807 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.049901 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.049983 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.050059 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:03Z","lastTransitionTime":"2025-12-09T16:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.120030 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.120087 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.120031 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:03 crc kubenswrapper[4954]: E1209 16:58:03.120194 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:03 crc kubenswrapper[4954]: E1209 16:58:03.120382 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:03 crc kubenswrapper[4954]: E1209 16:58:03.120516 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.152914 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.152955 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.152964 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.152979 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.152988 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:03Z","lastTransitionTime":"2025-12-09T16:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.255412 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.255452 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.255461 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.255477 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.255487 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:03Z","lastTransitionTime":"2025-12-09T16:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.357277 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.357519 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.357637 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.357738 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.357824 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:03Z","lastTransitionTime":"2025-12-09T16:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.460099 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.460126 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.460134 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.460148 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.460156 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:03Z","lastTransitionTime":"2025-12-09T16:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.562806 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.563091 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.563160 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.563223 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.563280 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:03Z","lastTransitionTime":"2025-12-09T16:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.666781 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.667453 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.667480 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.667519 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.667543 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:03Z","lastTransitionTime":"2025-12-09T16:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.770885 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.770952 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.770994 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.771018 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.771033 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:03Z","lastTransitionTime":"2025-12-09T16:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.873610 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.873663 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.873674 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.873694 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.873707 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:03Z","lastTransitionTime":"2025-12-09T16:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.976631 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.976697 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.976726 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.976759 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:03 crc kubenswrapper[4954]: I1209 16:58:03.976776 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:03Z","lastTransitionTime":"2025-12-09T16:58:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.079889 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.079999 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.080010 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.080026 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.080039 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:04Z","lastTransitionTime":"2025-12-09T16:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.119719 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:04 crc kubenswrapper[4954]: E1209 16:58:04.119908 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.139281 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6b37602ea166cf2129b2ab60447f820eea78e67cd1aecb51a600c398a2c020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.157705 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.181468 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5736742d-6a6c-4309-b95d-6103408901ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:58Z\\\",\\\"message\\\":\\\"n.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.88\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 16:57:58.387939 7006 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc in node crc\\\\nI1209 16:57:58.387945 7006 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc after 0 failed attempt(s)\\\\nI1209 16:57:58.387944 7006 services_controller.go:452] Built service openshift-console-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1209 16:57:58.387952 7006 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1209 16:57:58.387702 7006 ovn.go:134] Ensuring zone local for Pod openshift-kube-scheduler/openshift-kube-scheduler-crc in node crc\\\\nI1209 16:57:58.387959 7006 services_controller.go:453] Built service openshift-console-operator/metrics template LB for network=default: []services.LB{}\\\\nI1209 16:57:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msqm2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7b4bw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.182969 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.183014 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.183025 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.183043 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.183054 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:04Z","lastTransitionTime":"2025-12-09T16:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.198163 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"856fc314-24bd-403e-a4b0-3dcd73eba595\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf6cb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:16Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ffkzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.219088 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bf29ee1-e5f8-4027-b8c6-8e638a3218b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93550022c8e3fd7a490a48669ea5678f5c83d8ba7da19529cddad75fd2bc779f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://472d585e0d86c922a0612dcf3b18fb7909c4ca506b07ae91389195ba323e76e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://874e2b2e44ede914d776e4e212366644879d266cc7228c6477c86299bf989053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://741fcfdee22faf2a4a05744a563ab83b690f0a65623a4fcf17d84945b07b4a85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://617dc02d3db714784929222b6702882a87424322e5cb634bcd61a2b17ec9a32a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d3b835d467bd6360a77cf7c5699402db836da5ec666cb4d9016c4f39ede72d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://209537c9554b8cec81ce97b82786d12832165e459d5d36cf602dff93820ed60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://313679df140df101265c68ab7f9d012dc993356a3d58cae2da3ba9d857ef9d63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.234306 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e936476016b5e094c620aefa871df766133cdf4824f26417e5843662903704c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86cc605be420a5005eae2953497cedc63481e0201e72bd569edd33977ab82fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.249124 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9n2h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1eccf2e0-30a3-4201-a23f-85a92a522d72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T16:57:49Z\\\",\\\"message\\\":\\\"2025-12-09T16:57:04+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5\\\\n2025-12-09T16:57:04+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d1ccb502-a938-4048-be96-bcb17b5badd5 to /host/opt/cni/bin/\\\\n2025-12-09T16:57:04Z [verbose] multus-daemon started\\\\n2025-12-09T16:57:04Z [verbose] Readiness Indicator file check\\\\n2025-12-09T16:57:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7cs9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9n2h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.264069 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a6f4be9-a12a-4562-96b8-fd85694aa29c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3370ec62ae26ebb1ac16aa52156e325dfe9a443c247e58d53e2ece6f31ae6f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5f67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-clkc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.276632 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-r4gn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53c57bd8-d348-441c-b1b9-3e1e493d024c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8cd73c76536ea0ea829275f4096f8bb065a05e98003d00187a2c515cffdfb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrlqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-r4gn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.285897 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.286116 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.286199 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.286276 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.286344 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:04Z","lastTransitionTime":"2025-12-09T16:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.289196 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70175b95-63b3-485c-87a5-2e67167a7948\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e67bafbae7a178fc01850d99002644f862f46b393e7910931a884bd04a715bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b6201fcb70f6a14c1c11cfa20eddc0e05d563b276c2b41c06323e0fbb461945\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nrpkg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wxkr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.300771 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79d5afb1-2789-4058-97bf-b4604a9a5234\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17199e1bc8be0a9cb3e83dca2669f2d58187f64c1b82d3bf4a84b1a39538ec03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a710da94b9709afd9c0cd31bc88e2a7235cd21c360ae171c19bcc7f7210bed82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.314141 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e690f64-775b-4059-9819-18cc3c6da1c5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://327ec4b9b835a025b2efad8b30d5eab591d0860da7b3a214553b94b9b026c7f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e462de97284638e1161f14f470c25bad53981d0d08f81dde5980991e189edecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23844430c0b5f37c59d048d297bf804c4fbc96cddc4958465f909b24e2635e51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9c1439b10830990825f534081be6df6320f5649cb56b79755495f98e6c5ffbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.328252 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.347315 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb6c6493-2ebd-4f44-b988-2722f2f44eb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f29619122c84337338b928c7c890d9b83e4915858adeba00feed089d314e2e5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39d7ffa40d2726d0c0ae48797b5b18291b7a0e7605fd117c2b7f5139f29bb82d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5277a9751b4761c8f2db5c560465965a34e74f3d9773222253126a9d18c3ae91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c94e06d78c25ebc5cc6d877d444cf16f7adc79af6c8a0382d34a844fcf7dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d26cde29a34d1aae9e1b881491c8b2729833699bb317dca6d5dbbb29cc6befa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92ee88018a9b411c91828f0dbceae7b0647be447e3e13d9bd8367030f9a5ca40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713903f60878710cbf8bfb45187a521c2fe42679b707eaa79374599cb8d6b21c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:57:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:57:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nwdpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-x2fvb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.362521 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2a34813db440e6296f70d77ad200aa7ca24a1382e1c7e79f550152c3b79567\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.377233 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.389767 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p9vh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea104037-eb12-4e58-8313-b35f63b7aa3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78142f4896940b7357865f2a6725ad2d1e4efdb17e52972fb4b104a14f84187d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz6xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:57:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p9vh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.390100 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.390122 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.390133 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.390165 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.390179 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:04Z","lastTransitionTime":"2025-12-09T16:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.410478 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06b58b47-128b-4499-adcb-16e004b35592\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T16:57:03Z\\\",\\\"message\\\":\\\"file observer\\\\nW1209 16:57:02.294062 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1209 16:57:02.294338 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 16:57:02.295845 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-253772527/tls.crt::/tmp/serving-cert-253772527/tls.key\\\\\\\"\\\\nI1209 16:57:03.018323 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 16:57:03.021508 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 16:57:03.021538 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 16:57:03.021557 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 16:57:03.021562 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 16:57:03.029347 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 16:57:03.029379 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029385 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 16:57:03.029392 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 16:57:03.029396 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 16:57:03.029400 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 16:57:03.029404 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1209 16:57:03.029612 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1209 16:57:03.036015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:57:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T16:56:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.425451 4954 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9446d14f-2c37-49c4-9d37-cb3f54a24cbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:57:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T16:56:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fbc131d4fcf4303eae1be0f897cbf95e1cdba83ce17ee41d8b9705ee9a2e633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d56e2df978404aaf72a89b16b202cab13d0221370a549b5da43349999de909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72598a8f660ead4df8196e2356d6298a811eeaf2d65b8d260035d706ff148a17\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T16:56:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T16:56:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:04Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.492807 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.492872 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.492886 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.492907 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.493110 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:04Z","lastTransitionTime":"2025-12-09T16:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.595351 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.595399 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.595412 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.595428 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.595440 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:04Z","lastTransitionTime":"2025-12-09T16:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.699173 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.699243 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.699259 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.699286 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.699305 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:04Z","lastTransitionTime":"2025-12-09T16:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.801974 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.802023 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.802032 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.802050 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.802061 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:04Z","lastTransitionTime":"2025-12-09T16:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.904908 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.904953 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.904963 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.904982 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:04 crc kubenswrapper[4954]: I1209 16:58:04.904992 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:04Z","lastTransitionTime":"2025-12-09T16:58:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.007784 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.007814 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.007822 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.007837 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.007846 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:05Z","lastTransitionTime":"2025-12-09T16:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.111366 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.111419 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.111429 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.111448 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.111460 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:05Z","lastTransitionTime":"2025-12-09T16:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.119619 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.119650 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.119664 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:05 crc kubenswrapper[4954]: E1209 16:58:05.119763 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:05 crc kubenswrapper[4954]: E1209 16:58:05.119905 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:05 crc kubenswrapper[4954]: E1209 16:58:05.119964 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.214649 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.214703 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.214721 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.214746 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.214764 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:05Z","lastTransitionTime":"2025-12-09T16:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.317970 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.318048 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.318072 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.318107 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.318131 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:05Z","lastTransitionTime":"2025-12-09T16:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.420733 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.420805 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.420820 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.420838 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.420852 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:05Z","lastTransitionTime":"2025-12-09T16:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.523785 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.523825 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.523834 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.523848 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.523857 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:05Z","lastTransitionTime":"2025-12-09T16:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.626866 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.626952 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.626989 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.627023 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.627048 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:05Z","lastTransitionTime":"2025-12-09T16:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.730374 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.730443 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.730467 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.730496 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.730512 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:05Z","lastTransitionTime":"2025-12-09T16:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.832770 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.832812 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.832822 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.832840 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.832850 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:05Z","lastTransitionTime":"2025-12-09T16:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.935717 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.935760 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.935768 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.935787 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:05 crc kubenswrapper[4954]: I1209 16:58:05.935797 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:05Z","lastTransitionTime":"2025-12-09T16:58:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.037613 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.037688 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.037698 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.037711 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.037720 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:06Z","lastTransitionTime":"2025-12-09T16:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.119438 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:06 crc kubenswrapper[4954]: E1209 16:58:06.119548 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.139781 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.139823 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.139836 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.139854 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.139866 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:06Z","lastTransitionTime":"2025-12-09T16:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.243905 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.243958 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.243973 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.243999 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.244017 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:06Z","lastTransitionTime":"2025-12-09T16:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.347855 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.347906 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.347919 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.347944 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.347967 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:06Z","lastTransitionTime":"2025-12-09T16:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.450569 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.450677 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.450698 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.450733 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.450754 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:06Z","lastTransitionTime":"2025-12-09T16:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.554990 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.555120 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.555148 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.555190 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.555219 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:06Z","lastTransitionTime":"2025-12-09T16:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.657774 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.657829 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.657842 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.657858 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.657869 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:06Z","lastTransitionTime":"2025-12-09T16:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.761511 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.761559 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.761729 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.761753 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.761762 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:06Z","lastTransitionTime":"2025-12-09T16:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.865028 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.865084 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.865097 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.865117 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.865128 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:06Z","lastTransitionTime":"2025-12-09T16:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.968298 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.968357 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.968369 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.968388 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:06 crc kubenswrapper[4954]: I1209 16:58:06.968399 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:06Z","lastTransitionTime":"2025-12-09T16:58:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.011098 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.011268 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.011242364 +0000 UTC m=+147.399416194 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.070465 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.070524 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.070545 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.070565 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.070579 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.099277 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.099311 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.099319 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.099332 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.099341 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112273 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.112386 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.112441 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.112478 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.112524 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112610 4954 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112614 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112657 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112673 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.112654362 +0000 UTC m=+147.500828182 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112677 4954 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112710 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.112701513 +0000 UTC m=+147.500875333 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112623 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112734 4954 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112745 4954 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112616 4954 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112769 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.112761505 +0000 UTC m=+147.500935325 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.112812 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.112795356 +0000 UTC m=+147.500969196 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.116162 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.116196 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.116206 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.116223 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.116246 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.120138 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.120168 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.120306 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.120409 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.120540 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.120633 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.129947 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.134175 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.134234 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.134247 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.134269 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.134286 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.148303 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.151776 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.151808 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.151819 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.151835 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.151845 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.164502 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.168646 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.168696 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.168708 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.168727 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.168739 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.182788 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T16:58:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"936ce91d-f269-4212-970c-75535102a27e\\\",\\\"systemUUID\\\":\\\"d423a69c-3c1d-438f-aa6d-1e5bcc3d983e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T16:58:07Z is after 2025-08-24T17:21:41Z" Dec 09 16:58:07 crc kubenswrapper[4954]: E1209 16:58:07.182912 4954 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.184197 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.184226 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.184238 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.184254 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.184264 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.286941 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.286977 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.286988 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.287004 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.287015 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.389649 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.389704 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.389720 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.389742 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.389755 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.493045 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.493084 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.493096 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.493111 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.493120 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.596018 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.596060 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.596070 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.596087 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.596097 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.698484 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.698532 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.698544 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.698562 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.698574 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.800822 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.800850 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.800859 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.800874 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.800883 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.903803 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.903846 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.903856 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.903870 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:07 crc kubenswrapper[4954]: I1209 16:58:07.903879 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:07Z","lastTransitionTime":"2025-12-09T16:58:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.006105 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.006145 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.006156 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.006172 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.006184 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:08Z","lastTransitionTime":"2025-12-09T16:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.113123 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.113162 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.113171 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.113185 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.113194 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:08Z","lastTransitionTime":"2025-12-09T16:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.119355 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:08 crc kubenswrapper[4954]: E1209 16:58:08.119560 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.214830 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.214875 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.214886 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.214902 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.214912 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:08Z","lastTransitionTime":"2025-12-09T16:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.317117 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.317157 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.317167 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.317186 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.317205 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:08Z","lastTransitionTime":"2025-12-09T16:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.418803 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.418854 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.418865 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.418881 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.418890 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:08Z","lastTransitionTime":"2025-12-09T16:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.521920 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.521954 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.521965 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.521978 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.521989 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:08Z","lastTransitionTime":"2025-12-09T16:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.624106 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.624145 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.624156 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.624173 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.624211 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:08Z","lastTransitionTime":"2025-12-09T16:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.726752 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.726788 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.726797 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.726811 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.726821 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:08Z","lastTransitionTime":"2025-12-09T16:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.828962 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.829026 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.829037 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.829054 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.829065 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:08Z","lastTransitionTime":"2025-12-09T16:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.930827 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.930861 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.930870 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.930882 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:08 crc kubenswrapper[4954]: I1209 16:58:08.930891 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:08Z","lastTransitionTime":"2025-12-09T16:58:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.033585 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.033664 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.033674 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.033691 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.033703 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:09Z","lastTransitionTime":"2025-12-09T16:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.119708 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.119816 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.119827 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:09 crc kubenswrapper[4954]: E1209 16:58:09.120160 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:09 crc kubenswrapper[4954]: E1209 16:58:09.120305 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:09 crc kubenswrapper[4954]: E1209 16:58:09.120406 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.135888 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.135927 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.135938 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.135955 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.135966 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:09Z","lastTransitionTime":"2025-12-09T16:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.237669 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.237723 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.237734 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.237755 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.237768 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:09Z","lastTransitionTime":"2025-12-09T16:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.340710 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.340779 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.340791 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.340815 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.340836 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:09Z","lastTransitionTime":"2025-12-09T16:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.444375 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.444430 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.444442 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.444463 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.444476 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:09Z","lastTransitionTime":"2025-12-09T16:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.547073 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.547110 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.547126 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.547143 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.547152 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:09Z","lastTransitionTime":"2025-12-09T16:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.650526 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.650582 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.650608 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.650628 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.650639 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:09Z","lastTransitionTime":"2025-12-09T16:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.753434 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.753491 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.753502 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.753536 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.753551 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:09Z","lastTransitionTime":"2025-12-09T16:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.856492 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.856536 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.856563 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.856582 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.856617 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:09Z","lastTransitionTime":"2025-12-09T16:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.958883 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.958918 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.958941 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.958957 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:09 crc kubenswrapper[4954]: I1209 16:58:09.958966 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:09Z","lastTransitionTime":"2025-12-09T16:58:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.062097 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.062194 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.062241 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.062270 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.062286 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:10Z","lastTransitionTime":"2025-12-09T16:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.119919 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:10 crc kubenswrapper[4954]: E1209 16:58:10.120089 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.165236 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.165270 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.165280 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.165303 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.165316 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:10Z","lastTransitionTime":"2025-12-09T16:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.267767 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.267802 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.267812 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.267843 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.267852 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:10Z","lastTransitionTime":"2025-12-09T16:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.370483 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.370539 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.370553 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.370567 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.370577 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:10Z","lastTransitionTime":"2025-12-09T16:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.473098 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.473134 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.473146 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.473170 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.473183 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:10Z","lastTransitionTime":"2025-12-09T16:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.575203 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.575243 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.575255 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.575272 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.575282 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:10Z","lastTransitionTime":"2025-12-09T16:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.677844 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.677941 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.677976 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.678016 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.678039 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:10Z","lastTransitionTime":"2025-12-09T16:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.781831 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.781909 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.781934 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.781966 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.781987 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:10Z","lastTransitionTime":"2025-12-09T16:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.885822 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.885887 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.885905 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.885930 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.885944 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:10Z","lastTransitionTime":"2025-12-09T16:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.988506 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.988574 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.988625 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.988653 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:10 crc kubenswrapper[4954]: I1209 16:58:10.988674 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:10Z","lastTransitionTime":"2025-12-09T16:58:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.092630 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.092699 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.092726 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.092756 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.092777 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:11Z","lastTransitionTime":"2025-12-09T16:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.120078 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.120177 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.120248 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:11 crc kubenswrapper[4954]: E1209 16:58:11.120331 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:11 crc kubenswrapper[4954]: E1209 16:58:11.120499 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:11 crc kubenswrapper[4954]: E1209 16:58:11.120564 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.196039 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.196117 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.196140 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.196172 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.196198 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:11Z","lastTransitionTime":"2025-12-09T16:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.299481 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.299544 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.299561 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.299586 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.299638 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:11Z","lastTransitionTime":"2025-12-09T16:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.401805 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.401891 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.401916 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.401948 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.401971 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:11Z","lastTransitionTime":"2025-12-09T16:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.505001 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.505071 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.505093 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.505121 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.505144 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:11Z","lastTransitionTime":"2025-12-09T16:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.607880 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.607936 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.607951 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.607975 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.607993 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:11Z","lastTransitionTime":"2025-12-09T16:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.711229 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.711346 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.711377 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.711408 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.711433 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:11Z","lastTransitionTime":"2025-12-09T16:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.814039 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.814084 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.814119 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.814135 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.814146 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:11Z","lastTransitionTime":"2025-12-09T16:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.917325 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.917374 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.917385 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.917404 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:11 crc kubenswrapper[4954]: I1209 16:58:11.917414 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:11Z","lastTransitionTime":"2025-12-09T16:58:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.021280 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.021355 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.021373 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.021400 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.021417 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:12Z","lastTransitionTime":"2025-12-09T16:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.119634 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:12 crc kubenswrapper[4954]: E1209 16:58:12.119893 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.124675 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.124741 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.124750 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.124767 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.124778 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:12Z","lastTransitionTime":"2025-12-09T16:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.227164 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.227212 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.227226 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.227247 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.227262 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:12Z","lastTransitionTime":"2025-12-09T16:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.330815 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.330871 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.330888 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.330907 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.330919 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:12Z","lastTransitionTime":"2025-12-09T16:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.434265 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.434369 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.434406 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.434441 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.434466 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:12Z","lastTransitionTime":"2025-12-09T16:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.538052 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.538119 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.538134 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.538165 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.538182 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:12Z","lastTransitionTime":"2025-12-09T16:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.640993 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.641054 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.641076 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.641106 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.641119 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:12Z","lastTransitionTime":"2025-12-09T16:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.744858 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.744954 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.745012 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.745041 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.745092 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:12Z","lastTransitionTime":"2025-12-09T16:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.848568 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.848679 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.848731 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.848767 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.848855 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:12Z","lastTransitionTime":"2025-12-09T16:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.953845 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.953922 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.953943 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.953975 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:12 crc kubenswrapper[4954]: I1209 16:58:12.953998 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:12Z","lastTransitionTime":"2025-12-09T16:58:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.057287 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.057366 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.057380 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.057422 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.057438 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:13Z","lastTransitionTime":"2025-12-09T16:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.119271 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.119411 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:13 crc kubenswrapper[4954]: E1209 16:58:13.119532 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.119687 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:13 crc kubenswrapper[4954]: E1209 16:58:13.119737 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:13 crc kubenswrapper[4954]: E1209 16:58:13.120019 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.161193 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.161283 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.161306 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.161346 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.161371 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:13Z","lastTransitionTime":"2025-12-09T16:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.264865 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.264917 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.264927 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.264947 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.264958 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:13Z","lastTransitionTime":"2025-12-09T16:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.368578 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.368659 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.368672 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.368694 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.368708 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:13Z","lastTransitionTime":"2025-12-09T16:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.474468 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.474531 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.474548 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.474587 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.474624 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:13Z","lastTransitionTime":"2025-12-09T16:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.578001 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.578052 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.578068 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.578089 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.578103 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:13Z","lastTransitionTime":"2025-12-09T16:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.680521 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.680578 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.680615 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.680639 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.680652 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:13Z","lastTransitionTime":"2025-12-09T16:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.783428 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.783503 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.783532 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.783570 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.783636 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:13Z","lastTransitionTime":"2025-12-09T16:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.887732 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.887812 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.887831 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.887863 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.887882 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:13Z","lastTransitionTime":"2025-12-09T16:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.990241 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.990320 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.990332 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.990353 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:13 crc kubenswrapper[4954]: I1209 16:58:13.990364 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:13Z","lastTransitionTime":"2025-12-09T16:58:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.093633 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.093676 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.093690 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.093758 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.093772 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:14Z","lastTransitionTime":"2025-12-09T16:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.119770 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.120409 4954 scope.go:117] "RemoveContainer" containerID="6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3" Dec 09 16:58:14 crc kubenswrapper[4954]: E1209 16:58:14.120575 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" Dec 09 16:58:14 crc kubenswrapper[4954]: E1209 16:58:14.120862 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.197727 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.197784 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.197801 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.197828 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.197842 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:14Z","lastTransitionTime":"2025-12-09T16:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.206409 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-x2fvb" podStartSLOduration=72.206378638 podStartE2EDuration="1m12.206378638s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.20517256 +0000 UTC m=+90.593346380" watchObservedRunningTime="2025-12-09 16:58:14.206378638 +0000 UTC m=+90.594552478" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.219272 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.219257953 podStartE2EDuration="1m11.219257953s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.219098368 +0000 UTC m=+90.607272188" watchObservedRunningTime="2025-12-09 16:58:14.219257953 +0000 UTC m=+90.607431783" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.242137 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=67.242116262 podStartE2EDuration="1m7.242116262s" podCreationTimestamp="2025-12-09 16:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.231975393 +0000 UTC m=+90.620149213" watchObservedRunningTime="2025-12-09 16:58:14.242116262 +0000 UTC m=+90.630290082" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.276150 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-p9vh6" podStartSLOduration=72.276127991 podStartE2EDuration="1m12.276127991s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.242433042 +0000 UTC m=+90.630606862" watchObservedRunningTime="2025-12-09 16:58:14.276127991 +0000 UTC m=+90.664301811" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.276532 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=69.276529024 podStartE2EDuration="1m9.276529024s" podCreationTimestamp="2025-12-09 16:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.27579384 +0000 UTC m=+90.663967650" watchObservedRunningTime="2025-12-09 16:58:14.276529024 +0000 UTC m=+90.664702844" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.300372 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.300415 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.300429 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.300445 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.300456 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:14Z","lastTransitionTime":"2025-12-09T16:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.395695 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=40.395670759 podStartE2EDuration="40.395670759s" podCreationTimestamp="2025-12-09 16:57:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.39540601 +0000 UTC m=+90.783579820" watchObservedRunningTime="2025-12-09 16:58:14.395670759 +0000 UTC m=+90.783844589" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.395819 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=22.395814883 podStartE2EDuration="22.395814883s" podCreationTimestamp="2025-12-09 16:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.381957778 +0000 UTC m=+90.770131618" watchObservedRunningTime="2025-12-09 16:58:14.395814883 +0000 UTC m=+90.783988703" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.402850 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.402889 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.402898 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.402916 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.402927 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:14Z","lastTransitionTime":"2025-12-09T16:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.410186 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-h9n2h" podStartSLOduration=72.410170855 podStartE2EDuration="1m12.410170855s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.409789023 +0000 UTC m=+90.797962843" watchObservedRunningTime="2025-12-09 16:58:14.410170855 +0000 UTC m=+90.798344675" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.432164 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podStartSLOduration=72.432140125 podStartE2EDuration="1m12.432140125s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.421974166 +0000 UTC m=+90.810147986" watchObservedRunningTime="2025-12-09 16:58:14.432140125 +0000 UTC m=+90.820313945" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.433139 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-r4gn7" podStartSLOduration=72.433130127 podStartE2EDuration="1m12.433130127s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.432895339 +0000 UTC m=+90.821069159" watchObservedRunningTime="2025-12-09 16:58:14.433130127 +0000 UTC m=+90.821303947" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.505920 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.506001 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.506012 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.506030 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.506042 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:14Z","lastTransitionTime":"2025-12-09T16:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.608475 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.608553 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.608578 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.608647 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.608674 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:14Z","lastTransitionTime":"2025-12-09T16:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.711300 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.711391 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.711410 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.711438 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.711456 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:14Z","lastTransitionTime":"2025-12-09T16:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.814222 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.814280 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.814293 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.814314 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.814326 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:14Z","lastTransitionTime":"2025-12-09T16:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.916940 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.916987 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.916999 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.917021 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:14 crc kubenswrapper[4954]: I1209 16:58:14.917033 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:14Z","lastTransitionTime":"2025-12-09T16:58:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.020257 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.020312 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.020323 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.020340 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.020352 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:15Z","lastTransitionTime":"2025-12-09T16:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.119969 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:15 crc kubenswrapper[4954]: E1209 16:58:15.120105 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.120103 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.120155 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:15 crc kubenswrapper[4954]: E1209 16:58:15.120197 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:15 crc kubenswrapper[4954]: E1209 16:58:15.120381 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.123274 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.123307 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.123317 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.123334 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.123345 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:15Z","lastTransitionTime":"2025-12-09T16:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.225997 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.226058 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.226075 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.226098 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.226115 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:15Z","lastTransitionTime":"2025-12-09T16:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.329680 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.329779 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.329812 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.329847 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.329870 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:15Z","lastTransitionTime":"2025-12-09T16:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.433032 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.433098 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.433115 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.433139 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.433158 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:15Z","lastTransitionTime":"2025-12-09T16:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.535472 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.535527 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.535540 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.535562 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.535577 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:15Z","lastTransitionTime":"2025-12-09T16:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.638587 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.638716 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.638740 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.638774 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.638797 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:15Z","lastTransitionTime":"2025-12-09T16:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.742371 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.742421 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.742430 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.742448 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.742459 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:15Z","lastTransitionTime":"2025-12-09T16:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.845860 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.845932 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.845949 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.845974 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.845993 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:15Z","lastTransitionTime":"2025-12-09T16:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.949085 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.949152 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.949161 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.949178 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:15 crc kubenswrapper[4954]: I1209 16:58:15.949187 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:15Z","lastTransitionTime":"2025-12-09T16:58:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.051288 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.051328 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.051337 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.051351 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.051361 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:16Z","lastTransitionTime":"2025-12-09T16:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.119392 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:16 crc kubenswrapper[4954]: E1209 16:58:16.119508 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.153388 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.153440 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.153452 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.153470 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.153481 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:16Z","lastTransitionTime":"2025-12-09T16:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.255500 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.255541 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.255552 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.255568 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.255581 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:16Z","lastTransitionTime":"2025-12-09T16:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.357910 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.357943 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.357951 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.357965 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.357974 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:16Z","lastTransitionTime":"2025-12-09T16:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.460555 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.460610 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.460619 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.460635 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.460644 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:16Z","lastTransitionTime":"2025-12-09T16:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.563110 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.563167 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.563185 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.563203 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.563216 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:16Z","lastTransitionTime":"2025-12-09T16:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.665858 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.665942 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.665967 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.665999 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.666023 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:16Z","lastTransitionTime":"2025-12-09T16:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.768377 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.768443 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.768458 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.768476 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.768486 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:16Z","lastTransitionTime":"2025-12-09T16:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.871305 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.871340 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.871349 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.871361 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.871371 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:16Z","lastTransitionTime":"2025-12-09T16:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.974109 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.974165 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.974185 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.974210 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:16 crc kubenswrapper[4954]: I1209 16:58:16.974226 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:16Z","lastTransitionTime":"2025-12-09T16:58:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.077000 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.077059 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.077070 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.077090 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.077105 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:17Z","lastTransitionTime":"2025-12-09T16:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.119822 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.119931 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:17 crc kubenswrapper[4954]: E1209 16:58:17.120028 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.120049 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:17 crc kubenswrapper[4954]: E1209 16:58:17.120117 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:17 crc kubenswrapper[4954]: E1209 16:58:17.120298 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.180231 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.180303 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.180326 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.180356 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.180380 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:17Z","lastTransitionTime":"2025-12-09T16:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.283139 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.283241 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.283262 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.283286 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.283300 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:17Z","lastTransitionTime":"2025-12-09T16:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.386291 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.386325 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.386334 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.386349 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.386358 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:17Z","lastTransitionTime":"2025-12-09T16:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.392273 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.392323 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.392354 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.392374 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.392387 4954 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T16:58:17Z","lastTransitionTime":"2025-12-09T16:58:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.432282 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wxkr6" podStartSLOduration=74.432265162 podStartE2EDuration="1m14.432265162s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:14.448111618 +0000 UTC m=+90.836285438" watchObservedRunningTime="2025-12-09 16:58:17.432265162 +0000 UTC m=+93.820438982" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.432425 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc"] Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.432778 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.435573 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.435896 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.435977 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.437210 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.522407 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.522587 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.522692 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.522843 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.522883 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.623939 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.623995 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.624011 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.624050 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.624074 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.624073 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.624137 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.625448 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.631584 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.640291 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb7de78d-33d1-42a8-983e-6b4581b5e7b3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-tt8kc\" (UID: \"bb7de78d-33d1-42a8-983e-6b4581b5e7b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:17 crc kubenswrapper[4954]: I1209 16:58:17.750722 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" Dec 09 16:58:18 crc kubenswrapper[4954]: I1209 16:58:18.120029 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:18 crc kubenswrapper[4954]: E1209 16:58:18.120311 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:18 crc kubenswrapper[4954]: I1209 16:58:18.638962 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" event={"ID":"bb7de78d-33d1-42a8-983e-6b4581b5e7b3","Type":"ContainerStarted","Data":"631296a05b629bd5e3a2a1c0e3708912c570f0a9e4cbe622515249fd0810d219"} Dec 09 16:58:18 crc kubenswrapper[4954]: I1209 16:58:18.639064 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" event={"ID":"bb7de78d-33d1-42a8-983e-6b4581b5e7b3","Type":"ContainerStarted","Data":"8cbd7cb496037e563e3a578918fb1d805dd4af28039ecea5cd1353aeabae87a4"} Dec 09 16:58:18 crc kubenswrapper[4954]: I1209 16:58:18.654806 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tt8kc" podStartSLOduration=76.654783994 podStartE2EDuration="1m16.654783994s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:18.654376561 +0000 UTC m=+95.042550471" watchObservedRunningTime="2025-12-09 16:58:18.654783994 +0000 UTC m=+95.042957824" Dec 09 16:58:19 crc kubenswrapper[4954]: I1209 16:58:19.119966 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:19 crc kubenswrapper[4954]: I1209 16:58:19.120015 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:19 crc kubenswrapper[4954]: I1209 16:58:19.120013 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:19 crc kubenswrapper[4954]: E1209 16:58:19.120130 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:19 crc kubenswrapper[4954]: E1209 16:58:19.120197 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:19 crc kubenswrapper[4954]: E1209 16:58:19.120271 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:20 crc kubenswrapper[4954]: I1209 16:58:20.120118 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:20 crc kubenswrapper[4954]: E1209 16:58:20.120272 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:20 crc kubenswrapper[4954]: I1209 16:58:20.551741 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:20 crc kubenswrapper[4954]: E1209 16:58:20.551968 4954 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:58:20 crc kubenswrapper[4954]: E1209 16:58:20.552189 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs podName:856fc314-24bd-403e-a4b0-3dcd73eba595 nodeName:}" failed. No retries permitted until 2025-12-09 16:59:24.552167603 +0000 UTC m=+160.940341423 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs") pod "network-metrics-daemon-ffkzn" (UID: "856fc314-24bd-403e-a4b0-3dcd73eba595") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 16:58:21 crc kubenswrapper[4954]: I1209 16:58:21.119292 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:21 crc kubenswrapper[4954]: I1209 16:58:21.119355 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:21 crc kubenswrapper[4954]: I1209 16:58:21.119295 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:21 crc kubenswrapper[4954]: E1209 16:58:21.119427 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:21 crc kubenswrapper[4954]: E1209 16:58:21.119525 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:21 crc kubenswrapper[4954]: E1209 16:58:21.119653 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:22 crc kubenswrapper[4954]: I1209 16:58:22.119393 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:22 crc kubenswrapper[4954]: E1209 16:58:22.119673 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:23 crc kubenswrapper[4954]: I1209 16:58:23.120009 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:23 crc kubenswrapper[4954]: I1209 16:58:23.120109 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:23 crc kubenswrapper[4954]: E1209 16:58:23.120126 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:23 crc kubenswrapper[4954]: I1209 16:58:23.120317 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:23 crc kubenswrapper[4954]: E1209 16:58:23.120430 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:23 crc kubenswrapper[4954]: E1209 16:58:23.120664 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:24 crc kubenswrapper[4954]: I1209 16:58:24.119207 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:24 crc kubenswrapper[4954]: E1209 16:58:24.120644 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:25 crc kubenswrapper[4954]: I1209 16:58:25.120046 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:25 crc kubenswrapper[4954]: E1209 16:58:25.120160 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:25 crc kubenswrapper[4954]: I1209 16:58:25.120195 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:25 crc kubenswrapper[4954]: I1209 16:58:25.120201 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:25 crc kubenswrapper[4954]: E1209 16:58:25.120346 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:25 crc kubenswrapper[4954]: E1209 16:58:25.120779 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:25 crc kubenswrapper[4954]: I1209 16:58:25.121368 4954 scope.go:117] "RemoveContainer" containerID="6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3" Dec 09 16:58:25 crc kubenswrapper[4954]: E1209 16:58:25.121611 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" Dec 09 16:58:26 crc kubenswrapper[4954]: I1209 16:58:26.119639 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:26 crc kubenswrapper[4954]: E1209 16:58:26.119836 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:27 crc kubenswrapper[4954]: I1209 16:58:27.119900 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:27 crc kubenswrapper[4954]: I1209 16:58:27.119953 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:27 crc kubenswrapper[4954]: I1209 16:58:27.120004 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:27 crc kubenswrapper[4954]: E1209 16:58:27.120971 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:27 crc kubenswrapper[4954]: E1209 16:58:27.121119 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:27 crc kubenswrapper[4954]: E1209 16:58:27.120842 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:28 crc kubenswrapper[4954]: I1209 16:58:28.119888 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:28 crc kubenswrapper[4954]: E1209 16:58:28.120022 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:29 crc kubenswrapper[4954]: I1209 16:58:29.119572 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:29 crc kubenswrapper[4954]: I1209 16:58:29.119698 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:29 crc kubenswrapper[4954]: E1209 16:58:29.119724 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:29 crc kubenswrapper[4954]: E1209 16:58:29.119840 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:29 crc kubenswrapper[4954]: I1209 16:58:29.119898 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:29 crc kubenswrapper[4954]: E1209 16:58:29.119955 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:30 crc kubenswrapper[4954]: I1209 16:58:30.120522 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:30 crc kubenswrapper[4954]: E1209 16:58:30.120796 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:31 crc kubenswrapper[4954]: I1209 16:58:31.120153 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:31 crc kubenswrapper[4954]: I1209 16:58:31.120188 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:31 crc kubenswrapper[4954]: I1209 16:58:31.120415 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:31 crc kubenswrapper[4954]: E1209 16:58:31.120662 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:31 crc kubenswrapper[4954]: E1209 16:58:31.121110 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:31 crc kubenswrapper[4954]: E1209 16:58:31.121249 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:32 crc kubenswrapper[4954]: I1209 16:58:32.119488 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:32 crc kubenswrapper[4954]: E1209 16:58:32.119652 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:33 crc kubenswrapper[4954]: I1209 16:58:33.119379 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:33 crc kubenswrapper[4954]: I1209 16:58:33.119381 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:33 crc kubenswrapper[4954]: E1209 16:58:33.119496 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:33 crc kubenswrapper[4954]: I1209 16:58:33.119404 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:33 crc kubenswrapper[4954]: E1209 16:58:33.119543 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:33 crc kubenswrapper[4954]: E1209 16:58:33.120024 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:34 crc kubenswrapper[4954]: I1209 16:58:34.119438 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:34 crc kubenswrapper[4954]: E1209 16:58:34.120578 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:35 crc kubenswrapper[4954]: I1209 16:58:35.120065 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:35 crc kubenswrapper[4954]: I1209 16:58:35.120115 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:35 crc kubenswrapper[4954]: I1209 16:58:35.120209 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:35 crc kubenswrapper[4954]: E1209 16:58:35.120201 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:35 crc kubenswrapper[4954]: E1209 16:58:35.120325 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:35 crc kubenswrapper[4954]: E1209 16:58:35.120407 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:36 crc kubenswrapper[4954]: I1209 16:58:36.119302 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:36 crc kubenswrapper[4954]: E1209 16:58:36.119423 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:36 crc kubenswrapper[4954]: I1209 16:58:36.707571 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9n2h_1eccf2e0-30a3-4201-a23f-85a92a522d72/kube-multus/1.log" Dec 09 16:58:36 crc kubenswrapper[4954]: I1209 16:58:36.708025 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9n2h_1eccf2e0-30a3-4201-a23f-85a92a522d72/kube-multus/0.log" Dec 09 16:58:36 crc kubenswrapper[4954]: I1209 16:58:36.708074 4954 generic.go:334] "Generic (PLEG): container finished" podID="1eccf2e0-30a3-4201-a23f-85a92a522d72" containerID="adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010" exitCode=1 Dec 09 16:58:36 crc kubenswrapper[4954]: I1209 16:58:36.708109 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9n2h" event={"ID":"1eccf2e0-30a3-4201-a23f-85a92a522d72","Type":"ContainerDied","Data":"adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010"} Dec 09 16:58:36 crc kubenswrapper[4954]: I1209 16:58:36.708157 4954 scope.go:117] "RemoveContainer" containerID="3dbe67e4c51f5b6a76641b76ecadcb1f3ba6ee04f87f17723738e472fb57cb3c" Dec 09 16:58:36 crc kubenswrapper[4954]: I1209 16:58:36.708760 4954 scope.go:117] "RemoveContainer" containerID="adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010" Dec 09 16:58:36 crc kubenswrapper[4954]: E1209 16:58:36.709000 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-h9n2h_openshift-multus(1eccf2e0-30a3-4201-a23f-85a92a522d72)\"" pod="openshift-multus/multus-h9n2h" podUID="1eccf2e0-30a3-4201-a23f-85a92a522d72" Dec 09 16:58:37 crc kubenswrapper[4954]: I1209 16:58:37.119946 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:37 crc kubenswrapper[4954]: I1209 16:58:37.120025 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:37 crc kubenswrapper[4954]: I1209 16:58:37.120056 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:37 crc kubenswrapper[4954]: E1209 16:58:37.120176 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:37 crc kubenswrapper[4954]: E1209 16:58:37.120508 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:37 crc kubenswrapper[4954]: E1209 16:58:37.120695 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:37 crc kubenswrapper[4954]: I1209 16:58:37.120839 4954 scope.go:117] "RemoveContainer" containerID="6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3" Dec 09 16:58:37 crc kubenswrapper[4954]: E1209 16:58:37.121035 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7b4bw_openshift-ovn-kubernetes(5736742d-6a6c-4309-b95d-6103408901ec)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" Dec 09 16:58:37 crc kubenswrapper[4954]: I1209 16:58:37.713036 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9n2h_1eccf2e0-30a3-4201-a23f-85a92a522d72/kube-multus/1.log" Dec 09 16:58:38 crc kubenswrapper[4954]: I1209 16:58:38.119585 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:38 crc kubenswrapper[4954]: E1209 16:58:38.119733 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:39 crc kubenswrapper[4954]: I1209 16:58:39.119793 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:39 crc kubenswrapper[4954]: I1209 16:58:39.119850 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:39 crc kubenswrapper[4954]: I1209 16:58:39.120001 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:39 crc kubenswrapper[4954]: E1209 16:58:39.120128 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:39 crc kubenswrapper[4954]: E1209 16:58:39.120375 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:39 crc kubenswrapper[4954]: E1209 16:58:39.120430 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:40 crc kubenswrapper[4954]: I1209 16:58:40.120208 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:40 crc kubenswrapper[4954]: E1209 16:58:40.120451 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:41 crc kubenswrapper[4954]: I1209 16:58:41.119941 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:41 crc kubenswrapper[4954]: I1209 16:58:41.120031 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:41 crc kubenswrapper[4954]: E1209 16:58:41.120072 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:41 crc kubenswrapper[4954]: E1209 16:58:41.120194 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:41 crc kubenswrapper[4954]: I1209 16:58:41.120285 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:41 crc kubenswrapper[4954]: E1209 16:58:41.120460 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:42 crc kubenswrapper[4954]: I1209 16:58:42.119674 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:42 crc kubenswrapper[4954]: E1209 16:58:42.119795 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:43 crc kubenswrapper[4954]: I1209 16:58:43.119496 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:43 crc kubenswrapper[4954]: E1209 16:58:43.120174 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:43 crc kubenswrapper[4954]: I1209 16:58:43.119632 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:43 crc kubenswrapper[4954]: E1209 16:58:43.120395 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:43 crc kubenswrapper[4954]: I1209 16:58:43.119544 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:43 crc kubenswrapper[4954]: E1209 16:58:43.120576 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:44 crc kubenswrapper[4954]: I1209 16:58:44.119194 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:44 crc kubenswrapper[4954]: E1209 16:58:44.120446 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:44 crc kubenswrapper[4954]: E1209 16:58:44.142869 4954 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 09 16:58:44 crc kubenswrapper[4954]: E1209 16:58:44.258089 4954 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 16:58:45 crc kubenswrapper[4954]: I1209 16:58:45.119634 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:45 crc kubenswrapper[4954]: I1209 16:58:45.119702 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:45 crc kubenswrapper[4954]: E1209 16:58:45.119774 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:45 crc kubenswrapper[4954]: E1209 16:58:45.119948 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:45 crc kubenswrapper[4954]: I1209 16:58:45.120256 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:45 crc kubenswrapper[4954]: E1209 16:58:45.120384 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:46 crc kubenswrapper[4954]: I1209 16:58:46.119890 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:46 crc kubenswrapper[4954]: E1209 16:58:46.120125 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:47 crc kubenswrapper[4954]: I1209 16:58:47.119687 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:47 crc kubenswrapper[4954]: I1209 16:58:47.119765 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:47 crc kubenswrapper[4954]: E1209 16:58:47.119831 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:47 crc kubenswrapper[4954]: I1209 16:58:47.119710 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:47 crc kubenswrapper[4954]: E1209 16:58:47.119957 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:47 crc kubenswrapper[4954]: E1209 16:58:47.120130 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:48 crc kubenswrapper[4954]: I1209 16:58:48.119856 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:48 crc kubenswrapper[4954]: E1209 16:58:48.120015 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:49 crc kubenswrapper[4954]: I1209 16:58:49.120255 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:49 crc kubenswrapper[4954]: I1209 16:58:49.120299 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:49 crc kubenswrapper[4954]: I1209 16:58:49.120269 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:49 crc kubenswrapper[4954]: E1209 16:58:49.120422 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:49 crc kubenswrapper[4954]: E1209 16:58:49.120560 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:49 crc kubenswrapper[4954]: E1209 16:58:49.120717 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:49 crc kubenswrapper[4954]: E1209 16:58:49.259397 4954 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 16:58:50 crc kubenswrapper[4954]: I1209 16:58:50.119788 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:50 crc kubenswrapper[4954]: E1209 16:58:50.120127 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:51 crc kubenswrapper[4954]: I1209 16:58:51.119115 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:51 crc kubenswrapper[4954]: I1209 16:58:51.119150 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:51 crc kubenswrapper[4954]: E1209 16:58:51.119356 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:51 crc kubenswrapper[4954]: I1209 16:58:51.119369 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:51 crc kubenswrapper[4954]: E1209 16:58:51.119468 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:51 crc kubenswrapper[4954]: E1209 16:58:51.119555 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:51 crc kubenswrapper[4954]: I1209 16:58:51.119912 4954 scope.go:117] "RemoveContainer" containerID="adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010" Dec 09 16:58:51 crc kubenswrapper[4954]: I1209 16:58:51.760991 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9n2h_1eccf2e0-30a3-4201-a23f-85a92a522d72/kube-multus/1.log" Dec 09 16:58:51 crc kubenswrapper[4954]: I1209 16:58:51.761571 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9n2h" event={"ID":"1eccf2e0-30a3-4201-a23f-85a92a522d72","Type":"ContainerStarted","Data":"e0c232380ddfbb380b2f38c01d48754a185f0b2cee838436385d29e2945680e0"} Dec 09 16:58:52 crc kubenswrapper[4954]: I1209 16:58:52.119607 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:52 crc kubenswrapper[4954]: E1209 16:58:52.119721 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:52 crc kubenswrapper[4954]: I1209 16:58:52.120361 4954 scope.go:117] "RemoveContainer" containerID="6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3" Dec 09 16:58:52 crc kubenswrapper[4954]: I1209 16:58:52.767327 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/3.log" Dec 09 16:58:52 crc kubenswrapper[4954]: I1209 16:58:52.769986 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerStarted","Data":"03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc"} Dec 09 16:58:52 crc kubenswrapper[4954]: I1209 16:58:52.770749 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:58:52 crc kubenswrapper[4954]: I1209 16:58:52.914566 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podStartSLOduration=110.91453123 podStartE2EDuration="1m50.91453123s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:58:52.804154261 +0000 UTC m=+129.192328091" watchObservedRunningTime="2025-12-09 16:58:52.91453123 +0000 UTC m=+129.302705050" Dec 09 16:58:52 crc kubenswrapper[4954]: I1209 16:58:52.915661 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ffkzn"] Dec 09 16:58:52 crc kubenswrapper[4954]: I1209 16:58:52.915841 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:52 crc kubenswrapper[4954]: E1209 16:58:52.915989 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:53 crc kubenswrapper[4954]: I1209 16:58:53.120413 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:53 crc kubenswrapper[4954]: I1209 16:58:53.120491 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:53 crc kubenswrapper[4954]: E1209 16:58:53.121091 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:53 crc kubenswrapper[4954]: I1209 16:58:53.120515 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:53 crc kubenswrapper[4954]: E1209 16:58:53.121228 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:53 crc kubenswrapper[4954]: E1209 16:58:53.121270 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:54 crc kubenswrapper[4954]: E1209 16:58:54.260017 4954 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 16:58:55 crc kubenswrapper[4954]: I1209 16:58:55.120075 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:55 crc kubenswrapper[4954]: E1209 16:58:55.120227 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:55 crc kubenswrapper[4954]: I1209 16:58:55.120328 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:55 crc kubenswrapper[4954]: I1209 16:58:55.120448 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:55 crc kubenswrapper[4954]: E1209 16:58:55.120530 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:55 crc kubenswrapper[4954]: I1209 16:58:55.120683 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:55 crc kubenswrapper[4954]: E1209 16:58:55.120833 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:55 crc kubenswrapper[4954]: E1209 16:58:55.120996 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:57 crc kubenswrapper[4954]: I1209 16:58:57.120164 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:57 crc kubenswrapper[4954]: I1209 16:58:57.120180 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:57 crc kubenswrapper[4954]: E1209 16:58:57.120332 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:58:57 crc kubenswrapper[4954]: I1209 16:58:57.120231 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:57 crc kubenswrapper[4954]: E1209 16:58:57.120389 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:57 crc kubenswrapper[4954]: I1209 16:58:57.120217 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:57 crc kubenswrapper[4954]: E1209 16:58:57.120445 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:57 crc kubenswrapper[4954]: E1209 16:58:57.120509 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:59 crc kubenswrapper[4954]: I1209 16:58:59.119772 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:58:59 crc kubenswrapper[4954]: E1209 16:58:59.120508 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 16:58:59 crc kubenswrapper[4954]: I1209 16:58:59.119839 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:58:59 crc kubenswrapper[4954]: E1209 16:58:59.120748 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ffkzn" podUID="856fc314-24bd-403e-a4b0-3dcd73eba595" Dec 09 16:58:59 crc kubenswrapper[4954]: I1209 16:58:59.119813 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:58:59 crc kubenswrapper[4954]: I1209 16:58:59.119846 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:58:59 crc kubenswrapper[4954]: E1209 16:58:59.120912 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 16:58:59 crc kubenswrapper[4954]: E1209 16:58:59.121033 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 16:59:01 crc kubenswrapper[4954]: I1209 16:59:01.119738 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:59:01 crc kubenswrapper[4954]: I1209 16:59:01.119810 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:59:01 crc kubenswrapper[4954]: I1209 16:59:01.119826 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:59:01 crc kubenswrapper[4954]: I1209 16:59:01.119936 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:59:01 crc kubenswrapper[4954]: I1209 16:59:01.122722 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 09 16:59:01 crc kubenswrapper[4954]: I1209 16:59:01.122722 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 09 16:59:01 crc kubenswrapper[4954]: I1209 16:59:01.122921 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 09 16:59:01 crc kubenswrapper[4954]: I1209 16:59:01.122947 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 09 16:59:01 crc kubenswrapper[4954]: I1209 16:59:01.122942 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 09 16:59:01 crc kubenswrapper[4954]: I1209 16:59:01.130580 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.231676 4954 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.266302 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.266766 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.266934 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.267158 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-v6ftv"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.267379 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.267995 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.269036 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6449l"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.269287 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.269679 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.270076 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.270612 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.275789 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.277329 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.294585 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h7bjt"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.294702 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.294749 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.294795 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.294816 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.294758 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.295022 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.295181 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.295219 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.295500 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.295694 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.296012 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-v86j2"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.296396 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.296653 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.296773 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.297197 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.297252 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.297755 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.298264 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.298886 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.298988 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.299105 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.299146 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.299285 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.299447 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.299540 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.299622 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.299699 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.299771 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.299919 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.299950 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.300012 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.300109 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.300133 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.300236 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.301461 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.301543 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nq2tq"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.302080 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.302473 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-v4pm7"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.302818 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.305119 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.305258 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6gmvm"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.305799 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.306266 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.306614 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.306764 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.306267 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.306508 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.306775 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.307175 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.306868 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.307481 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.311048 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.311359 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dx5bf"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.311721 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.312015 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.312349 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.314121 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.327357 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.329368 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-96v5v"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.351540 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.351588 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.351865 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.351966 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.352164 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.352388 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.332095 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.352744 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.353097 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.353284 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.353525 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.353695 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.353998 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.354381 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.354673 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.354838 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.355020 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.357045 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.357302 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.357301 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.357554 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.357740 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.358057 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.359924 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.360110 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.360253 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.360361 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.360479 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.360586 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.360730 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.365329 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mhkmt"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.365808 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-cj74l"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.366147 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.366379 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.366798 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.367083 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.374334 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.375210 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.379182 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.379633 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.388243 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.388526 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.400006 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.400715 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pz4hd"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.401095 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8jkvw"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.401442 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.401729 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.401882 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.404671 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qmwzp"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.405400 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.410375 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-service-ca\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.412859 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413233 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1de74a6-26b3-46df-b9bf-717509ae4150-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413319 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c84a7cef-2031-4477-adbb-0e5f97cf7726-config\") pod \"openshift-apiserver-operator-796bbdcf4f-w2spv\" (UID: \"c84a7cef-2031-4477-adbb-0e5f97cf7726\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413349 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-trusted-ca\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413380 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a770208-d33d-48d7-b70c-be193682fb72-serving-cert\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413413 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/db03f79f-fcea-4a92-b021-bac0b6c1cebc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qpkjl\" (UID: \"db03f79f-fcea-4a92-b021-bac0b6c1cebc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413456 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-serving-cert\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413550 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413561 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413623 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s6bz\" (UniqueName: \"kubernetes.io/projected/6a770208-d33d-48d7-b70c-be193682fb72-kube-api-access-6s6bz\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413648 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-serving-cert\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413671 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjm9q\" (UniqueName: \"kubernetes.io/projected/dc11e3bb-97da-485b-ad28-34b0116ed220-kube-api-access-kjm9q\") pod \"downloads-7954f5f757-6gmvm\" (UID: \"dc11e3bb-97da-485b-ad28-34b0116ed220\") " pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413695 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad6309f7-1627-4f65-9169-181e7a046adc-config\") pod \"kube-apiserver-operator-766d6c64bb-gcjjz\" (UID: \"ad6309f7-1627-4f65-9169-181e7a046adc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413724 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdpsk\" (UniqueName: \"kubernetes.io/projected/a759536e-62a7-4410-b1a3-0ea059cd63f4-kube-api-access-sdpsk\") pod \"dns-operator-744455d44c-96v5v\" (UID: \"a759536e-62a7-4410-b1a3-0ea059cd63f4\") " pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413747 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413770 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1de74a6-26b3-46df-b9bf-717509ae4150-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413793 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmv2s\" (UniqueName: \"kubernetes.io/projected/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-kube-api-access-wmv2s\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413816 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvx6z\" (UniqueName: \"kubernetes.io/projected/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-kube-api-access-tvx6z\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413842 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-trusted-ca-bundle\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413865 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-oauth-serving-cert\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.413905 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrzdf\" (UniqueName: \"kubernetes.io/projected/db03f79f-fcea-4a92-b021-bac0b6c1cebc-kube-api-access-xrzdf\") pod \"openshift-config-operator-7777fb866f-qpkjl\" (UID: \"db03f79f-fcea-4a92-b021-bac0b6c1cebc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.419571 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxcgq\" (UniqueName: \"kubernetes.io/projected/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-kube-api-access-jxcgq\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.419685 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.419738 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-config\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.419791 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c84a7cef-2031-4477-adbb-0e5f97cf7726-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-w2spv\" (UID: \"c84a7cef-2031-4477-adbb-0e5f97cf7726\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.419855 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvdlz\" (UniqueName: \"kubernetes.io/projected/82cdcdae-1e7b-4591-80f7-b297bef6d91f-kube-api-access-dvdlz\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.419887 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-config\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.419937 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-client-ca\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420013 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420047 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420041 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-audit-policies\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420104 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420183 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29zbk\" (UniqueName: \"kubernetes.io/projected/a1de74a6-26b3-46df-b9bf-717509ae4150-kube-api-access-29zbk\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420226 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-etcd-client\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420254 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-encryption-config\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420308 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad6309f7-1627-4f65-9169-181e7a046adc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gcjjz\" (UID: \"ad6309f7-1627-4f65-9169-181e7a046adc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420353 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad6309f7-1627-4f65-9169-181e7a046adc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gcjjz\" (UID: \"ad6309f7-1627-4f65-9169-181e7a046adc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420413 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-config\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420479 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1de74a6-26b3-46df-b9bf-717509ae4150-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420511 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-config\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420535 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9wtr\" (UniqueName: \"kubernetes.io/projected/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-kube-api-access-d9wtr\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420560 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phgqt\" (UniqueName: \"kubernetes.io/projected/4a1f21d3-3aee-45dc-baa6-2929b5f6e150-kube-api-access-phgqt\") pod \"migrator-59844c95c7-5dhxr\" (UID: \"4a1f21d3-3aee-45dc-baa6-2929b5f6e150\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420626 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jxtz\" (UniqueName: \"kubernetes.io/projected/cd86b184-0ad7-416c-a326-70dad6e39296-kube-api-access-9jxtz\") pod \"cluster-samples-operator-665b6dd947-48vd4\" (UID: \"cd86b184-0ad7-416c-a326-70dad6e39296\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420672 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kfq8\" (UniqueName: \"kubernetes.io/projected/c84a7cef-2031-4477-adbb-0e5f97cf7726-kube-api-access-5kfq8\") pod \"openshift-apiserver-operator-796bbdcf4f-w2spv\" (UID: \"c84a7cef-2031-4477-adbb-0e5f97cf7726\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420707 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-serving-cert\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420730 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd86b184-0ad7-416c-a326-70dad6e39296-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-48vd4\" (UID: \"cd86b184-0ad7-416c-a326-70dad6e39296\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420789 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-audit-dir\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420845 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/82cdcdae-1e7b-4591-80f7-b297bef6d91f-machine-approver-tls\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420893 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-oauth-config\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420925 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-images\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420973 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db03f79f-fcea-4a92-b021-bac0b6c1cebc-serving-cert\") pod \"openshift-config-operator-7777fb866f-qpkjl\" (UID: \"db03f79f-fcea-4a92-b021-bac0b6c1cebc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.420999 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a759536e-62a7-4410-b1a3-0ea059cd63f4-metrics-tls\") pod \"dns-operator-744455d44c-96v5v\" (UID: \"a759536e-62a7-4410-b1a3-0ea059cd63f4\") " pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.421061 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/82cdcdae-1e7b-4591-80f7-b297bef6d91f-auth-proxy-config\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.421097 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82cdcdae-1e7b-4591-80f7-b297bef6d91f-config\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.422927 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.425276 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.425811 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.439714 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.440085 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.444189 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nslc6"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.445093 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.445456 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.446145 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.448775 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.449472 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.449955 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.450957 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.452294 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.452729 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.453021 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.453166 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.453284 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.453401 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.453656 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.453776 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.454096 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.454184 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.454242 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.454295 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.454257 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.454395 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.454490 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.454548 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.454695 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.455012 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.455141 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.455630 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.455735 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.455871 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.456116 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.456416 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.456702 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.457042 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.457368 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.457720 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.457875 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.457985 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.459475 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.460280 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.478248 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.478893 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.484764 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.488206 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mnc4f"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.498304 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.499751 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.502789 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.504641 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.504920 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.505327 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.508232 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.509344 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.510881 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.511197 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-v6ftv"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.512410 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.512662 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.513644 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6449l"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.514948 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h7bjt"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.518570 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-cmrfr"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.519699 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cmrfr" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524449 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrzdf\" (UniqueName: \"kubernetes.io/projected/db03f79f-fcea-4a92-b021-bac0b6c1cebc-kube-api-access-xrzdf\") pod \"openshift-config-operator-7777fb866f-qpkjl\" (UID: \"db03f79f-fcea-4a92-b021-bac0b6c1cebc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524493 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxcgq\" (UniqueName: \"kubernetes.io/projected/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-kube-api-access-jxcgq\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524534 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-config\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524560 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c84a7cef-2031-4477-adbb-0e5f97cf7726-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-w2spv\" (UID: \"c84a7cef-2031-4477-adbb-0e5f97cf7726\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524586 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvdlz\" (UniqueName: \"kubernetes.io/projected/82cdcdae-1e7b-4591-80f7-b297bef6d91f-kube-api-access-dvdlz\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524628 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-config\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524649 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-client-ca\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524672 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-audit-policies\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524696 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524717 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29zbk\" (UniqueName: \"kubernetes.io/projected/a1de74a6-26b3-46df-b9bf-717509ae4150-kube-api-access-29zbk\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524737 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-etcd-client\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524757 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-encryption-config\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524778 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad6309f7-1627-4f65-9169-181e7a046adc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gcjjz\" (UID: \"ad6309f7-1627-4f65-9169-181e7a046adc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524821 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad6309f7-1627-4f65-9169-181e7a046adc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gcjjz\" (UID: \"ad6309f7-1627-4f65-9169-181e7a046adc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524845 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-config\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524880 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1de74a6-26b3-46df-b9bf-717509ae4150-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524904 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-config\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524927 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9wtr\" (UniqueName: \"kubernetes.io/projected/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-kube-api-access-d9wtr\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524949 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phgqt\" (UniqueName: \"kubernetes.io/projected/4a1f21d3-3aee-45dc-baa6-2929b5f6e150-kube-api-access-phgqt\") pod \"migrator-59844c95c7-5dhxr\" (UID: \"4a1f21d3-3aee-45dc-baa6-2929b5f6e150\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.524975 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jxtz\" (UniqueName: \"kubernetes.io/projected/cd86b184-0ad7-416c-a326-70dad6e39296-kube-api-access-9jxtz\") pod \"cluster-samples-operator-665b6dd947-48vd4\" (UID: \"cd86b184-0ad7-416c-a326-70dad6e39296\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525000 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kfq8\" (UniqueName: \"kubernetes.io/projected/c84a7cef-2031-4477-adbb-0e5f97cf7726-kube-api-access-5kfq8\") pod \"openshift-apiserver-operator-796bbdcf4f-w2spv\" (UID: \"c84a7cef-2031-4477-adbb-0e5f97cf7726\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525023 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-serving-cert\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525050 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd86b184-0ad7-416c-a326-70dad6e39296-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-48vd4\" (UID: \"cd86b184-0ad7-416c-a326-70dad6e39296\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525074 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-audit-dir\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525099 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/82cdcdae-1e7b-4591-80f7-b297bef6d91f-machine-approver-tls\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525123 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-oauth-config\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525146 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-images\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525175 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db03f79f-fcea-4a92-b021-bac0b6c1cebc-serving-cert\") pod \"openshift-config-operator-7777fb866f-qpkjl\" (UID: \"db03f79f-fcea-4a92-b021-bac0b6c1cebc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525201 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a759536e-62a7-4410-b1a3-0ea059cd63f4-metrics-tls\") pod \"dns-operator-744455d44c-96v5v\" (UID: \"a759536e-62a7-4410-b1a3-0ea059cd63f4\") " pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525235 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/82cdcdae-1e7b-4591-80f7-b297bef6d91f-auth-proxy-config\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525258 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82cdcdae-1e7b-4591-80f7-b297bef6d91f-config\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525288 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-service-ca\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525310 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1de74a6-26b3-46df-b9bf-717509ae4150-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525331 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c84a7cef-2031-4477-adbb-0e5f97cf7726-config\") pod \"openshift-apiserver-operator-796bbdcf4f-w2spv\" (UID: \"c84a7cef-2031-4477-adbb-0e5f97cf7726\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525354 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-trusted-ca\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525377 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a770208-d33d-48d7-b70c-be193682fb72-serving-cert\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525400 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/db03f79f-fcea-4a92-b021-bac0b6c1cebc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qpkjl\" (UID: \"db03f79f-fcea-4a92-b021-bac0b6c1cebc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525423 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-serving-cert\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525447 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525494 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s6bz\" (UniqueName: \"kubernetes.io/projected/6a770208-d33d-48d7-b70c-be193682fb72-kube-api-access-6s6bz\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525517 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-serving-cert\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525541 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjm9q\" (UniqueName: \"kubernetes.io/projected/dc11e3bb-97da-485b-ad28-34b0116ed220-kube-api-access-kjm9q\") pod \"downloads-7954f5f757-6gmvm\" (UID: \"dc11e3bb-97da-485b-ad28-34b0116ed220\") " pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525567 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad6309f7-1627-4f65-9169-181e7a046adc-config\") pod \"kube-apiserver-operator-766d6c64bb-gcjjz\" (UID: \"ad6309f7-1627-4f65-9169-181e7a046adc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525612 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdpsk\" (UniqueName: \"kubernetes.io/projected/a759536e-62a7-4410-b1a3-0ea059cd63f4-kube-api-access-sdpsk\") pod \"dns-operator-744455d44c-96v5v\" (UID: \"a759536e-62a7-4410-b1a3-0ea059cd63f4\") " pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525640 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525663 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1de74a6-26b3-46df-b9bf-717509ae4150-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525688 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmv2s\" (UniqueName: \"kubernetes.io/projected/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-kube-api-access-wmv2s\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525715 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvx6z\" (UniqueName: \"kubernetes.io/projected/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-kube-api-access-tvx6z\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525738 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-trusted-ca-bundle\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525737 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.525762 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-oauth-serving-cert\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.526534 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.526737 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-oauth-serving-cert\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.531236 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-config\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.532609 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-service-ca\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.533029 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-trusted-ca\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.533068 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.534198 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/82cdcdae-1e7b-4591-80f7-b297bef6d91f-machine-approver-tls\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.534384 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-images\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.534994 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/db03f79f-fcea-4a92-b021-bac0b6c1cebc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-qpkjl\" (UID: \"db03f79f-fcea-4a92-b021-bac0b6c1cebc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.535801 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c84a7cef-2031-4477-adbb-0e5f97cf7726-config\") pod \"openshift-apiserver-operator-796bbdcf4f-w2spv\" (UID: \"c84a7cef-2031-4477-adbb-0e5f97cf7726\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.537505 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-config\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.538344 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-audit-policies\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.538993 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.539112 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1de74a6-26b3-46df-b9bf-717509ae4150-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.540219 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/82cdcdae-1e7b-4591-80f7-b297bef6d91f-auth-proxy-config\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.540356 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-config\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.540523 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-serving-cert\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.541024 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.542836 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-client-ca\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.543240 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-trusted-ca-bundle\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.543727 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-audit-dir\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.544148 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82cdcdae-1e7b-4591-80f7-b297bef6d91f-config\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.544952 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-config\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.545122 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a770208-d33d-48d7-b70c-be193682fb72-serving-cert\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.546352 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-v86j2"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.548698 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-serving-cert\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.549082 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-oauth-config\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.550815 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.554705 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.556304 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8jkvw"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.562069 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a759536e-62a7-4410-b1a3-0ea059cd63f4-metrics-tls\") pod \"dns-operator-744455d44c-96v5v\" (UID: \"a759536e-62a7-4410-b1a3-0ea059cd63f4\") " pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.562847 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c84a7cef-2031-4477-adbb-0e5f97cf7726-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-w2spv\" (UID: \"c84a7cef-2031-4477-adbb-0e5f97cf7726\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.564204 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1de74a6-26b3-46df-b9bf-717509ae4150-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.564377 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd86b184-0ad7-416c-a326-70dad6e39296-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-48vd4\" (UID: \"cd86b184-0ad7-416c-a326-70dad6e39296\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.567939 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad6309f7-1627-4f65-9169-181e7a046adc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gcjjz\" (UID: \"ad6309f7-1627-4f65-9169-181e7a046adc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.568382 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.569087 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.570094 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.570154 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.570418 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-encryption-config\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.570823 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-etcd-client\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.570822 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db03f79f-fcea-4a92-b021-bac0b6c1cebc-serving-cert\") pod \"openshift-config-operator-7777fb866f-qpkjl\" (UID: \"db03f79f-fcea-4a92-b021-bac0b6c1cebc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.570896 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6gmvm"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.571233 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-serving-cert\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.573332 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pz4hd"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.574572 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-96v5v"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.575523 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad6309f7-1627-4f65-9169-181e7a046adc-config\") pod \"kube-apiserver-operator-766d6c64bb-gcjjz\" (UID: \"ad6309f7-1627-4f65-9169-181e7a046adc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.575587 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dx5bf"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.576670 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.580393 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-gw877"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.581501 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.581523 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nq2tq"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.581617 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gw877" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.583754 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-fvjws"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.584496 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.586628 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.588027 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mhkmt"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.593260 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.594912 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nslc6"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.596485 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qmwzp"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.600952 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mnc4f"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.602857 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.606401 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.609243 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.611472 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.619445 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gw877"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.621826 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-v4pm7"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.626699 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.628944 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.631399 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-b4gqt"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.632873 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.633707 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.635330 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.636479 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cmrfr"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.638029 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.639490 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.641348 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.645012 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-b4gqt"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.646504 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9"] Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.649107 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.669261 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.690306 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.709668 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.730246 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.749471 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.769072 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.789400 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.809791 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.830058 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.849379 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.870656 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.889690 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.910180 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.929630 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.949916 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.969715 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 09 16:59:08 crc kubenswrapper[4954]: I1209 16:59:08.990506 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.009642 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.029170 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.050468 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.070392 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.089431 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.110023 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.130162 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.148904 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.170086 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.190031 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.209728 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.230074 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.250271 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.276419 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.289525 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.309446 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.329286 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.349082 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.369034 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.389074 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.407486 4954 request.go:700] Waited for 1.001688246s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.408891 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.428999 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.449272 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.469168 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.489408 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.508801 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.529394 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.548992 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.569977 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.589490 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.608968 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.628667 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.648710 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.669520 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.689189 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.709632 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.736250 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.749486 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.789389 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.809452 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.828647 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.848827 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.868460 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.889096 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.908998 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.934156 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.949282 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.970528 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 09 16:59:09 crc kubenswrapper[4954]: I1209 16:59:09.989257 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.010246 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.029790 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.049664 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.070184 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.090559 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.108887 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.128779 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.166553 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrzdf\" (UniqueName: \"kubernetes.io/projected/db03f79f-fcea-4a92-b021-bac0b6c1cebc-kube-api-access-xrzdf\") pod \"openshift-config-operator-7777fb866f-qpkjl\" (UID: \"db03f79f-fcea-4a92-b021-bac0b6c1cebc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.183341 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxcgq\" (UniqueName: \"kubernetes.io/projected/8ffdbd23-8e68-4bf3-bb37-2817fa11f18b-kube-api-access-jxcgq\") pod \"machine-api-operator-5694c8668f-h7bjt\" (UID: \"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.202885 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdpsk\" (UniqueName: \"kubernetes.io/projected/a759536e-62a7-4410-b1a3-0ea059cd63f4-kube-api-access-sdpsk\") pod \"dns-operator-744455d44c-96v5v\" (UID: \"a759536e-62a7-4410-b1a3-0ea059cd63f4\") " pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.221502 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1de74a6-26b3-46df-b9bf-717509ae4150-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.249496 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmv2s\" (UniqueName: \"kubernetes.io/projected/73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe-kube-api-access-wmv2s\") pod \"console-operator-58897d9998-v86j2\" (UID: \"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe\") " pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.254058 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.263028 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvx6z\" (UniqueName: \"kubernetes.io/projected/ff5d80e6-ad76-49c5-a350-c7cd9e8528ae-kube-api-access-tvx6z\") pod \"apiserver-7bbb656c7d-hp66s\" (UID: \"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.284134 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjm9q\" (UniqueName: \"kubernetes.io/projected/dc11e3bb-97da-485b-ad28-34b0116ed220-kube-api-access-kjm9q\") pod \"downloads-7954f5f757-6gmvm\" (UID: \"dc11e3bb-97da-485b-ad28-34b0116ed220\") " pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.305870 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ad6309f7-1627-4f65-9169-181e7a046adc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gcjjz\" (UID: \"ad6309f7-1627-4f65-9169-181e7a046adc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.328751 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.338797 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvdlz\" (UniqueName: \"kubernetes.io/projected/82cdcdae-1e7b-4591-80f7-b297bef6d91f-kube-api-access-dvdlz\") pod \"machine-approver-56656f9798-nzwb5\" (UID: \"82cdcdae-1e7b-4591-80f7-b297bef6d91f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.350222 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29zbk\" (UniqueName: \"kubernetes.io/projected/a1de74a6-26b3-46df-b9bf-717509ae4150-kube-api-access-29zbk\") pod \"cluster-image-registry-operator-dc59b4c8b-hfd9b\" (UID: \"a1de74a6-26b3-46df-b9bf-717509ae4150\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.364510 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jxtz\" (UniqueName: \"kubernetes.io/projected/cd86b184-0ad7-416c-a326-70dad6e39296-kube-api-access-9jxtz\") pod \"cluster-samples-operator-665b6dd947-48vd4\" (UID: \"cd86b184-0ad7-416c-a326-70dad6e39296\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.388016 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.390296 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kfq8\" (UniqueName: \"kubernetes.io/projected/c84a7cef-2031-4477-adbb-0e5f97cf7726-kube-api-access-5kfq8\") pod \"openshift-apiserver-operator-796bbdcf4f-w2spv\" (UID: \"c84a7cef-2031-4477-adbb-0e5f97cf7726\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.406787 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.415259 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phgqt\" (UniqueName: \"kubernetes.io/projected/4a1f21d3-3aee-45dc-baa6-2929b5f6e150-kube-api-access-phgqt\") pod \"migrator-59844c95c7-5dhxr\" (UID: \"4a1f21d3-3aee-45dc-baa6-2929b5f6e150\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.415410 4954 request.go:700] Waited for 1.871682637s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa/token Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.428914 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.431502 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.437740 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.440685 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s6bz\" (UniqueName: \"kubernetes.io/projected/6a770208-d33d-48d7-b70c-be193682fb72-kube-api-access-6s6bz\") pod \"route-controller-manager-6576b87f9c-5cftp\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.441289 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h7bjt"] Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.447637 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.448766 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9wtr\" (UniqueName: \"kubernetes.io/projected/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-kube-api-access-d9wtr\") pod \"console-f9d7485db-v4pm7\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.449634 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.457108 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.470647 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.477970 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.489983 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.510148 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.513144 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.523762 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s"] Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.527008 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.529152 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.535898 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.550438 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.590164 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.610829 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.613852 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.630675 4954 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651019 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-config\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651064 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651095 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n8gl\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-kube-api-access-9n8gl\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651120 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e43b0681-33e3-4eca-a737-ed3c9c2fe994-audit-dir\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651143 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhkx9\" (UniqueName: \"kubernetes.io/projected/9ada019a-2740-482e-864b-55762c330b26-kube-api-access-rhkx9\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651202 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e43b0681-33e3-4eca-a737-ed3c9c2fe994-encryption-config\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651232 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651258 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-trusted-ca\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651324 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1623d36d-33ed-4c91-9069-aba484b54a83-config\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651481 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4jct\" (UniqueName: \"kubernetes.io/projected/1df4a2c0-8c86-447c-aa38-cb9a436f84eb-kube-api-access-r4jct\") pod \"openshift-controller-manager-operator-756b6f6bc6-j9crx\" (UID: \"1df4a2c0-8c86-447c-aa38-cb9a436f84eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651579 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e43b0681-33e3-4eca-a737-ed3c9c2fe994-node-pullsecrets\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651682 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57dc805d-5ae9-4f59-80fb-0dd9e15a33e0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8gzxc\" (UID: \"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651708 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9ada019a-2740-482e-864b-55762c330b26-audit-dir\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651733 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1623d36d-33ed-4c91-9069-aba484b54a83-service-ca-bundle\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651772 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57dc805d-5ae9-4f59-80fb-0dd9e15a33e0-proxy-tls\") pod \"machine-config-controller-84d6567774-8gzxc\" (UID: \"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651799 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651870 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-certificates\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651925 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651957 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.651988 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652018 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d284ca9d-999e-4ab5-8d7e-647af799d69c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652043 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-audit-policies\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652072 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfrvc\" (UniqueName: \"kubernetes.io/projected/57dc805d-5ae9-4f59-80fb-0dd9e15a33e0-kube-api-access-nfrvc\") pod \"machine-config-controller-84d6567774-8gzxc\" (UID: \"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652095 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652136 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1df4a2c0-8c86-447c-aa38-cb9a436f84eb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-j9crx\" (UID: \"1df4a2c0-8c86-447c-aa38-cb9a436f84eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652158 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-image-import-ca\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652187 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652244 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652266 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652295 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-etcd-serving-ca\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652321 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-bound-sa-token\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652351 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d284ca9d-999e-4ab5-8d7e-647af799d69c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652385 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x8w8\" (UniqueName: \"kubernetes.io/projected/e43b0681-33e3-4eca-a737-ed3c9c2fe994-kube-api-access-5x8w8\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652406 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1df4a2c0-8c86-447c-aa38-cb9a436f84eb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-j9crx\" (UID: \"1df4a2c0-8c86-447c-aa38-cb9a436f84eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652433 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e43b0681-33e3-4eca-a737-ed3c9c2fe994-serving-cert\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652473 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-audit\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652490 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e43b0681-33e3-4eca-a737-ed3c9c2fe994-etcd-client\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652509 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652536 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1623d36d-33ed-4c91-9069-aba484b54a83-serving-cert\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652568 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-tls\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652584 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-trusted-ca-bundle\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652621 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nm9j\" (UniqueName: \"kubernetes.io/projected/1623d36d-33ed-4c91-9069-aba484b54a83-kube-api-access-7nm9j\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652648 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.652668 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1623d36d-33ed-4c91-9069-aba484b54a83-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: E1209 16:59:10.653853 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.153835151 +0000 UTC m=+147.542008971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.753646 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.753857 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d284ca9d-999e-4ab5-8d7e-647af799d69c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.753891 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mldc6\" (UniqueName: \"kubernetes.io/projected/ffebb621-9aa6-4104-a457-c180e6c93a28-kube-api-access-mldc6\") pod \"service-ca-9c57cc56f-qmwzp\" (UID: \"ffebb621-9aa6-4104-a457-c180e6c93a28\") " pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.753914 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jch89\" (UniqueName: \"kubernetes.io/projected/afae1fb2-041c-433d-9e7b-e29a7e82ed31-kube-api-access-jch89\") pod \"package-server-manager-789f6589d5-cbfp9\" (UID: \"afae1fb2-041c-433d-9e7b-e29a7e82ed31\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.753945 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x8w8\" (UniqueName: \"kubernetes.io/projected/e43b0681-33e3-4eca-a737-ed3c9c2fe994-kube-api-access-5x8w8\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.753968 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec10355-df85-4f79-9bd5-78ff480e25f4-serving-cert\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.753982 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-mountpoint-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.754121 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b423076b-47af-4e4e-a085-d47074f6bb8f-cert\") pod \"ingress-canary-cmrfr\" (UID: \"b423076b-47af-4e4e-a085-d47074f6bb8f\") " pod="openshift-ingress-canary/ingress-canary-cmrfr" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.754143 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1df4a2c0-8c86-447c-aa38-cb9a436f84eb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-j9crx\" (UID: \"1df4a2c0-8c86-447c-aa38-cb9a436f84eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.754162 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e43b0681-33e3-4eca-a737-ed3c9c2fe994-serving-cert\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.754409 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eee9658e-289f-4ecf-8661-5b1bdf9c307c-config-volume\") pod \"dns-default-gw877\" (UID: \"eee9658e-289f-4ecf-8661-5b1bdf9c307c\") " pod="openshift-dns/dns-default-gw877" Dec 09 16:59:10 crc kubenswrapper[4954]: E1209 16:59:10.754652 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.254573671 +0000 UTC m=+147.642747631 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.755523 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-audit\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.755627 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-config\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.755924 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dd6facda-112d-4218-9cec-3e24d3d2249d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.755958 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/afae1fb2-041c-433d-9e7b-e29a7e82ed31-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cbfp9\" (UID: \"afae1fb2-041c-433d-9e7b-e29a7e82ed31\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.755998 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-tls\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756018 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-trusted-ca-bundle\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756132 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd7jd\" (UniqueName: \"kubernetes.io/projected/dd6facda-112d-4218-9cec-3e24d3d2249d-kube-api-access-dd7jd\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756246 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-audit\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756252 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nm9j\" (UniqueName: \"kubernetes.io/projected/1623d36d-33ed-4c91-9069-aba484b54a83-kube-api-access-7nm9j\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756421 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-config\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756475 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f82b145c-6847-4a03-afff-2a513a79ecbe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-prclt\" (UID: \"f82b145c-6847-4a03-afff-2a513a79ecbe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756497 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n8gl\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-kube-api-access-9n8gl\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756516 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e43b0681-33e3-4eca-a737-ed3c9c2fe994-audit-dir\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756534 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhkx9\" (UniqueName: \"kubernetes.io/projected/9ada019a-2740-482e-864b-55762c330b26-kube-api-access-rhkx9\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756586 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8jkvw\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756633 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e43b0681-33e3-4eca-a737-ed3c9c2fe994-encryption-config\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756663 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-trusted-ca\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756679 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-csi-data-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756694 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-config-volume\") pod \"collect-profiles-29421645-tk9h9\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756709 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2pvl\" (UniqueName: \"kubernetes.io/projected/cbebf2b4-7952-45d6-8c77-efc6438e05fd-kube-api-access-g2pvl\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756772 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2qm8\" (UniqueName: \"kubernetes.io/projected/fcc2c29a-2e97-457a-8f25-d96a4aee4ea5-kube-api-access-s2qm8\") pod \"multus-admission-controller-857f4d67dd-pz4hd\" (UID: \"fcc2c29a-2e97-457a-8f25-d96a4aee4ea5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756801 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bll4m\" (UniqueName: \"kubernetes.io/projected/849c8fd6-0873-40d0-8c40-7bbc2365452d-kube-api-access-bll4m\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756826 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8jkvw\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756850 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-secret-volume\") pod \"collect-profiles-29421645-tk9h9\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756866 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cbebf2b4-7952-45d6-8c77-efc6438e05fd-stats-auth\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756887 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4jct\" (UniqueName: \"kubernetes.io/projected/1df4a2c0-8c86-447c-aa38-cb9a436f84eb-kube-api-access-r4jct\") pod \"openshift-controller-manager-operator-756b6f6bc6-j9crx\" (UID: \"1df4a2c0-8c86-447c-aa38-cb9a436f84eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756910 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e43b0681-33e3-4eca-a737-ed3c9c2fe994-node-pullsecrets\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756970 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d30d8b55-366e-44df-967d-57e72d253fe6-certs\") pod \"machine-config-server-fvjws\" (UID: \"d30d8b55-366e-44df-967d-57e72d253fe6\") " pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.756999 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df27667c-ecf7-4e33-9d34-c3a5b4088154-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-q4bd2\" (UID: \"df27667c-ecf7-4e33-9d34-c3a5b4088154\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757007 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-trusted-ca-bundle\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757040 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df27667c-ecf7-4e33-9d34-c3a5b4088154-config\") pod \"kube-controller-manager-operator-78b949d7b-q4bd2\" (UID: \"df27667c-ecf7-4e33-9d34-c3a5b4088154\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757062 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9ada019a-2740-482e-864b-55762c330b26-audit-dir\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757082 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1623d36d-33ed-4c91-9069-aba484b54a83-service-ca-bundle\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757100 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4e939827-530f-4e2e-9455-080369e5c76d-profile-collector-cert\") pod \"catalog-operator-68c6474976-r7zd5\" (UID: \"4e939827-530f-4e2e-9455-080369e5c76d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757116 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cbebf2b4-7952-45d6-8c77-efc6438e05fd-metrics-certs\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757141 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57dc805d-5ae9-4f59-80fb-0dd9e15a33e0-proxy-tls\") pod \"machine-config-controller-84d6567774-8gzxc\" (UID: \"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757158 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757190 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-client-ca\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757208 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3ae04268-6e5b-402f-b3c3-6b456df7e689-webhook-cert\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757311 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec10355-df85-4f79-9bd5-78ff480e25f4-config\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757339 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-socket-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757360 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/849c8fd6-0873-40d0-8c40-7bbc2365452d-metrics-tls\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757376 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05f2d085-d179-42d9-bc77-7dce694be634-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zlqrk\" (UID: \"05f2d085-d179-42d9-bc77-7dce694be634\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757392 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwphl\" (UniqueName: \"kubernetes.io/projected/f82b145c-6847-4a03-afff-2a513a79ecbe-kube-api-access-kwphl\") pod \"kube-storage-version-migrator-operator-b67b599dd-prclt\" (UID: \"f82b145c-6847-4a03-afff-2a513a79ecbe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757408 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f82b145c-6847-4a03-afff-2a513a79ecbe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-prclt\" (UID: \"f82b145c-6847-4a03-afff-2a513a79ecbe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757425 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x52dh\" (UniqueName: \"kubernetes.io/projected/192c8b71-0927-4047-8fa6-a0b45fdbef7b-kube-api-access-x52dh\") pod \"olm-operator-6b444d44fb-sspwb\" (UID: \"192c8b71-0927-4047-8fa6-a0b45fdbef7b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757453 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d284ca9d-999e-4ab5-8d7e-647af799d69c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757472 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-audit-policies\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757488 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7ec10355-df85-4f79-9bd5-78ff480e25f4-etcd-client\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757503 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4brz\" (UniqueName: \"kubernetes.io/projected/7ec10355-df85-4f79-9bd5-78ff480e25f4-kube-api-access-q4brz\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757519 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnqqf\" (UniqueName: \"kubernetes.io/projected/454e49f6-b054-4901-bd18-461d7e9cd89e-kube-api-access-rnqqf\") pod \"marketplace-operator-79b997595-8jkvw\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757560 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757580 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8rxv\" (UniqueName: \"kubernetes.io/projected/eee9658e-289f-4ecf-8661-5b1bdf9c307c-kube-api-access-n8rxv\") pod \"dns-default-gw877\" (UID: \"eee9658e-289f-4ecf-8661-5b1bdf9c307c\") " pod="openshift-dns/dns-default-gw877" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757623 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757640 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cbebf2b4-7952-45d6-8c77-efc6438e05fd-default-certificate\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757668 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d30d8b55-366e-44df-967d-57e72d253fe6-node-bootstrap-token\") pod \"machine-config-server-fvjws\" (UID: \"d30d8b55-366e-44df-967d-57e72d253fe6\") " pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757716 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr8c9\" (UniqueName: \"kubernetes.io/projected/b423076b-47af-4e4e-a085-d47074f6bb8f-kube-api-access-hr8c9\") pod \"ingress-canary-cmrfr\" (UID: \"b423076b-47af-4e4e-a085-d47074f6bb8f\") " pod="openshift-ingress-canary/ingress-canary-cmrfr" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757927 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05f2d085-d179-42d9-bc77-7dce694be634-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zlqrk\" (UID: \"05f2d085-d179-42d9-bc77-7dce694be634\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757953 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757972 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9de03415-ecf6-43a2-8968-eae338108559-serving-cert\") pod \"service-ca-operator-777779d784-nslc6\" (UID: \"9de03415-ecf6-43a2-8968-eae338108559\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.757993 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-bound-sa-token\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758009 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7ec10355-df85-4f79-9bd5-78ff480e25f4-etcd-ca\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758027 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ffebb621-9aa6-4104-a457-c180e6c93a28-signing-cabundle\") pod \"service-ca-9c57cc56f-qmwzp\" (UID: \"ffebb621-9aa6-4104-a457-c180e6c93a28\") " pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758043 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbebf2b4-7952-45d6-8c77-efc6438e05fd-service-ca-bundle\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758059 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrwq6\" (UniqueName: \"kubernetes.io/projected/3ae04268-6e5b-402f-b3c3-6b456df7e689-kube-api-access-mrwq6\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758089 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-plugins-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758106 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9de03415-ecf6-43a2-8968-eae338108559-config\") pod \"service-ca-operator-777779d784-nslc6\" (UID: \"9de03415-ecf6-43a2-8968-eae338108559\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758147 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e43b0681-33e3-4eca-a737-ed3c9c2fe994-etcd-client\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758164 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4e939827-530f-4e2e-9455-080369e5c76d-srv-cert\") pod \"catalog-operator-68c6474976-r7zd5\" (UID: \"4e939827-530f-4e2e-9455-080369e5c76d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758180 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/849c8fd6-0873-40d0-8c40-7bbc2365452d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758198 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f80de45-61f5-4538-93d7-c6252e3c64bc-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7jpc4\" (UID: \"8f80de45-61f5-4538-93d7-c6252e3c64bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758223 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758241 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ffebb621-9aa6-4104-a457-c180e6c93a28-signing-key\") pod \"service-ca-9c57cc56f-qmwzp\" (UID: \"ffebb621-9aa6-4104-a457-c180e6c93a28\") " pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758257 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7n2c\" (UniqueName: \"kubernetes.io/projected/1cb535ff-3daa-418e-929e-99664eea736b-kube-api-access-r7n2c\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758257 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-trusted-ca\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758273 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/192c8b71-0927-4047-8fa6-a0b45fdbef7b-srv-cert\") pod \"olm-operator-6b444d44fb-sspwb\" (UID: \"192c8b71-0927-4047-8fa6-a0b45fdbef7b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758367 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1623d36d-33ed-4c91-9069-aba484b54a83-serving-cert\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758398 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdfkq\" (UniqueName: \"kubernetes.io/projected/4fc4931a-8c26-4f29-8798-d730e10e8d39-kube-api-access-sdfkq\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758418 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4zvs\" (UniqueName: \"kubernetes.io/projected/9de03415-ecf6-43a2-8968-eae338108559-kube-api-access-n4zvs\") pod \"service-ca-operator-777779d784-nslc6\" (UID: \"9de03415-ecf6-43a2-8968-eae338108559\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758462 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758480 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1623d36d-33ed-4c91-9069-aba484b54a83-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758498 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/eee9658e-289f-4ecf-8661-5b1bdf9c307c-metrics-tls\") pod \"dns-default-gw877\" (UID: \"eee9658e-289f-4ecf-8661-5b1bdf9c307c\") " pod="openshift-dns/dns-default-gw877" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758539 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758581 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758614 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cb535ff-3daa-418e-929e-99664eea736b-serving-cert\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758765 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1623d36d-33ed-4c91-9069-aba484b54a83-config\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758841 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3ae04268-6e5b-402f-b3c3-6b456df7e689-tmpfs\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758898 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dd6facda-112d-4218-9cec-3e24d3d2249d-proxy-tls\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758944 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57dc805d-5ae9-4f59-80fb-0dd9e15a33e0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8gzxc\" (UID: \"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.758965 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759012 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-registration-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759034 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dd6facda-112d-4218-9cec-3e24d3d2249d-images\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759057 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fcc2c29a-2e97-457a-8f25-d96a4aee4ea5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pz4hd\" (UID: \"fcc2c29a-2e97-457a-8f25-d96a4aee4ea5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759079 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3ae04268-6e5b-402f-b3c3-6b456df7e689-apiservice-cert\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759191 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-certificates\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759215 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bfpg\" (UniqueName: \"kubernetes.io/projected/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-kube-api-access-6bfpg\") pod \"collect-profiles-29421645-tk9h9\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759240 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bspcd\" (UniqueName: \"kubernetes.io/projected/4e939827-530f-4e2e-9455-080369e5c76d-kube-api-access-bspcd\") pod \"catalog-operator-68c6474976-r7zd5\" (UID: \"4e939827-530f-4e2e-9455-080369e5c76d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759356 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df27667c-ecf7-4e33-9d34-c3a5b4088154-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-q4bd2\" (UID: \"df27667c-ecf7-4e33-9d34-c3a5b4088154\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759387 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/05f2d085-d179-42d9-bc77-7dce694be634-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zlqrk\" (UID: \"05f2d085-d179-42d9-bc77-7dce694be634\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759455 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759517 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759539 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ec10355-df85-4f79-9bd5-78ff480e25f4-etcd-service-ca\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759558 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g28g9\" (UniqueName: \"kubernetes.io/projected/8f80de45-61f5-4538-93d7-c6252e3c64bc-kube-api-access-g28g9\") pod \"control-plane-machine-set-operator-78cbb6b69f-7jpc4\" (UID: \"8f80de45-61f5-4538-93d7-c6252e3c64bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759575 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcr8d\" (UniqueName: \"kubernetes.io/projected/d30d8b55-366e-44df-967d-57e72d253fe6-kube-api-access-jcr8d\") pod \"machine-config-server-fvjws\" (UID: \"d30d8b55-366e-44df-967d-57e72d253fe6\") " pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759627 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759671 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/849c8fd6-0873-40d0-8c40-7bbc2365452d-trusted-ca\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759700 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfrvc\" (UniqueName: \"kubernetes.io/projected/57dc805d-5ae9-4f59-80fb-0dd9e15a33e0-kube-api-access-nfrvc\") pod \"machine-config-controller-84d6567774-8gzxc\" (UID: \"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759729 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1df4a2c0-8c86-447c-aa38-cb9a436f84eb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-j9crx\" (UID: \"1df4a2c0-8c86-447c-aa38-cb9a436f84eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.759750 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-image-import-ca\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.760311 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-etcd-serving-ca\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.760335 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.760353 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/192c8b71-0927-4047-8fa6-a0b45fdbef7b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sspwb\" (UID: \"192c8b71-0927-4047-8fa6-a0b45fdbef7b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:10 crc kubenswrapper[4954]: E1209 16:59:10.760811 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.260792118 +0000 UTC m=+147.648965938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.761185 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e43b0681-33e3-4eca-a737-ed3c9c2fe994-audit-dir\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.761326 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d284ca9d-999e-4ab5-8d7e-647af799d69c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.761690 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e43b0681-33e3-4eca-a737-ed3c9c2fe994-node-pullsecrets\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.762503 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.766097 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.766300 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-config\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.766509 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e43b0681-33e3-4eca-a737-ed3c9c2fe994-etcd-client\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.766525 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-image-import-ca\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.767179 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d284ca9d-999e-4ab5-8d7e-647af799d69c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.767567 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1623d36d-33ed-4c91-9069-aba484b54a83-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.769037 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9ada019a-2740-482e-864b-55762c330b26-audit-dir\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.769529 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1623d36d-33ed-4c91-9069-aba484b54a83-service-ca-bundle\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.769761 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/57dc805d-5ae9-4f59-80fb-0dd9e15a33e0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8gzxc\" (UID: \"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.770565 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-tls\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.771359 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/57dc805d-5ae9-4f59-80fb-0dd9e15a33e0-proxy-tls\") pod \"machine-config-controller-84d6567774-8gzxc\" (UID: \"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.772677 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-audit-policies\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.773357 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.776740 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.777313 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e43b0681-33e3-4eca-a737-ed3c9c2fe994-serving-cert\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.778533 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.779346 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-certificates\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.780373 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.780497 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.780931 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1df4a2c0-8c86-447c-aa38-cb9a436f84eb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-j9crx\" (UID: \"1df4a2c0-8c86-447c-aa38-cb9a436f84eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.781720 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1df4a2c0-8c86-447c-aa38-cb9a436f84eb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-j9crx\" (UID: \"1df4a2c0-8c86-447c-aa38-cb9a436f84eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.782364 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1623d36d-33ed-4c91-9069-aba484b54a83-config\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.782453 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e43b0681-33e3-4eca-a737-ed3c9c2fe994-encryption-config\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.782742 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.782945 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.783402 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e43b0681-33e3-4eca-a737-ed3c9c2fe994-etcd-serving-ca\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.788200 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1623d36d-33ed-4c91-9069-aba484b54a83-serving-cert\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.791617 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.817735 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x8w8\" (UniqueName: \"kubernetes.io/projected/e43b0681-33e3-4eca-a737-ed3c9c2fe994-kube-api-access-5x8w8\") pod \"apiserver-76f77b778f-v6ftv\" (UID: \"e43b0681-33e3-4eca-a737-ed3c9c2fe994\") " pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.826064 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.829173 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nm9j\" (UniqueName: \"kubernetes.io/projected/1623d36d-33ed-4c91-9069-aba484b54a83-kube-api-access-7nm9j\") pod \"authentication-operator-69f744f599-nq2tq\" (UID: \"1623d36d-33ed-4c91-9069-aba484b54a83\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.831361 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" event={"ID":"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b","Type":"ContainerStarted","Data":"01d76aa3838526203fec48db6853121743dd151d50e8a6a0262ca04a8ae66408"} Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.831418 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" event={"ID":"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b","Type":"ContainerStarted","Data":"cf9612b3049214b1a5313262513030bf93e8e6e076d0881d11cc12443d8d95b9"} Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.833200 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" event={"ID":"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae","Type":"ContainerStarted","Data":"54b57fb23e591940e335efb4c168815a2770ec06fbb6cbd2270c2fb18d2c9671"} Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.835302 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" event={"ID":"82cdcdae-1e7b-4591-80f7-b297bef6d91f","Type":"ContainerStarted","Data":"f7ff24443453229cd218bc416f03ad89909680858e025ef28b8067e4b5351a71"} Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.853002 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n8gl\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-kube-api-access-9n8gl\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861157 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:10 crc kubenswrapper[4954]: E1209 16:59:10.861255 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.361231406 +0000 UTC m=+147.749405226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861486 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd7jd\" (UniqueName: \"kubernetes.io/projected/dd6facda-112d-4218-9cec-3e24d3d2249d-kube-api-access-dd7jd\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861516 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f82b145c-6847-4a03-afff-2a513a79ecbe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-prclt\" (UID: \"f82b145c-6847-4a03-afff-2a513a79ecbe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861541 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8jkvw\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861559 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2pvl\" (UniqueName: \"kubernetes.io/projected/cbebf2b4-7952-45d6-8c77-efc6438e05fd-kube-api-access-g2pvl\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861578 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2qm8\" (UniqueName: \"kubernetes.io/projected/fcc2c29a-2e97-457a-8f25-d96a4aee4ea5-kube-api-access-s2qm8\") pod \"multus-admission-controller-857f4d67dd-pz4hd\" (UID: \"fcc2c29a-2e97-457a-8f25-d96a4aee4ea5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861616 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-csi-data-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861635 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-config-volume\") pod \"collect-profiles-29421645-tk9h9\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861652 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bll4m\" (UniqueName: \"kubernetes.io/projected/849c8fd6-0873-40d0-8c40-7bbc2365452d-kube-api-access-bll4m\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861673 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8jkvw\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861690 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-secret-volume\") pod \"collect-profiles-29421645-tk9h9\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861705 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cbebf2b4-7952-45d6-8c77-efc6438e05fd-stats-auth\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861727 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d30d8b55-366e-44df-967d-57e72d253fe6-certs\") pod \"machine-config-server-fvjws\" (UID: \"d30d8b55-366e-44df-967d-57e72d253fe6\") " pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861747 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df27667c-ecf7-4e33-9d34-c3a5b4088154-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-q4bd2\" (UID: \"df27667c-ecf7-4e33-9d34-c3a5b4088154\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861763 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df27667c-ecf7-4e33-9d34-c3a5b4088154-config\") pod \"kube-controller-manager-operator-78b949d7b-q4bd2\" (UID: \"df27667c-ecf7-4e33-9d34-c3a5b4088154\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861782 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4e939827-530f-4e2e-9455-080369e5c76d-profile-collector-cert\") pod \"catalog-operator-68c6474976-r7zd5\" (UID: \"4e939827-530f-4e2e-9455-080369e5c76d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861805 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cbebf2b4-7952-45d6-8c77-efc6438e05fd-metrics-certs\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861828 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-client-ca\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861845 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3ae04268-6e5b-402f-b3c3-6b456df7e689-webhook-cert\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861861 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec10355-df85-4f79-9bd5-78ff480e25f4-config\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861878 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-socket-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861893 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/849c8fd6-0873-40d0-8c40-7bbc2365452d-metrics-tls\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861908 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05f2d085-d179-42d9-bc77-7dce694be634-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zlqrk\" (UID: \"05f2d085-d179-42d9-bc77-7dce694be634\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861924 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwphl\" (UniqueName: \"kubernetes.io/projected/f82b145c-6847-4a03-afff-2a513a79ecbe-kube-api-access-kwphl\") pod \"kube-storage-version-migrator-operator-b67b599dd-prclt\" (UID: \"f82b145c-6847-4a03-afff-2a513a79ecbe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861938 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f82b145c-6847-4a03-afff-2a513a79ecbe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-prclt\" (UID: \"f82b145c-6847-4a03-afff-2a513a79ecbe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861957 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x52dh\" (UniqueName: \"kubernetes.io/projected/192c8b71-0927-4047-8fa6-a0b45fdbef7b-kube-api-access-x52dh\") pod \"olm-operator-6b444d44fb-sspwb\" (UID: \"192c8b71-0927-4047-8fa6-a0b45fdbef7b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861973 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnqqf\" (UniqueName: \"kubernetes.io/projected/454e49f6-b054-4901-bd18-461d7e9cd89e-kube-api-access-rnqqf\") pod \"marketplace-operator-79b997595-8jkvw\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.861990 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7ec10355-df85-4f79-9bd5-78ff480e25f4-etcd-client\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862009 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4brz\" (UniqueName: \"kubernetes.io/projected/7ec10355-df85-4f79-9bd5-78ff480e25f4-kube-api-access-q4brz\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862024 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8rxv\" (UniqueName: \"kubernetes.io/projected/eee9658e-289f-4ecf-8661-5b1bdf9c307c-kube-api-access-n8rxv\") pod \"dns-default-gw877\" (UID: \"eee9658e-289f-4ecf-8661-5b1bdf9c307c\") " pod="openshift-dns/dns-default-gw877" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862046 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cbebf2b4-7952-45d6-8c77-efc6438e05fd-default-certificate\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862064 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d30d8b55-366e-44df-967d-57e72d253fe6-node-bootstrap-token\") pod \"machine-config-server-fvjws\" (UID: \"d30d8b55-366e-44df-967d-57e72d253fe6\") " pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862082 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr8c9\" (UniqueName: \"kubernetes.io/projected/b423076b-47af-4e4e-a085-d47074f6bb8f-kube-api-access-hr8c9\") pod \"ingress-canary-cmrfr\" (UID: \"b423076b-47af-4e4e-a085-d47074f6bb8f\") " pod="openshift-ingress-canary/ingress-canary-cmrfr" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862100 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05f2d085-d179-42d9-bc77-7dce694be634-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zlqrk\" (UID: \"05f2d085-d179-42d9-bc77-7dce694be634\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862119 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9de03415-ecf6-43a2-8968-eae338108559-serving-cert\") pod \"service-ca-operator-777779d784-nslc6\" (UID: \"9de03415-ecf6-43a2-8968-eae338108559\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862136 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbebf2b4-7952-45d6-8c77-efc6438e05fd-service-ca-bundle\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862156 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrwq6\" (UniqueName: \"kubernetes.io/projected/3ae04268-6e5b-402f-b3c3-6b456df7e689-kube-api-access-mrwq6\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862177 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7ec10355-df85-4f79-9bd5-78ff480e25f4-etcd-ca\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862197 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ffebb621-9aa6-4104-a457-c180e6c93a28-signing-cabundle\") pod \"service-ca-9c57cc56f-qmwzp\" (UID: \"ffebb621-9aa6-4104-a457-c180e6c93a28\") " pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862210 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9de03415-ecf6-43a2-8968-eae338108559-config\") pod \"service-ca-operator-777779d784-nslc6\" (UID: \"9de03415-ecf6-43a2-8968-eae338108559\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862226 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-plugins-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862252 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4e939827-530f-4e2e-9455-080369e5c76d-srv-cert\") pod \"catalog-operator-68c6474976-r7zd5\" (UID: \"4e939827-530f-4e2e-9455-080369e5c76d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862265 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/849c8fd6-0873-40d0-8c40-7bbc2365452d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862283 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f80de45-61f5-4538-93d7-c6252e3c64bc-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7jpc4\" (UID: \"8f80de45-61f5-4538-93d7-c6252e3c64bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862302 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ffebb621-9aa6-4104-a457-c180e6c93a28-signing-key\") pod \"service-ca-9c57cc56f-qmwzp\" (UID: \"ffebb621-9aa6-4104-a457-c180e6c93a28\") " pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862318 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7n2c\" (UniqueName: \"kubernetes.io/projected/1cb535ff-3daa-418e-929e-99664eea736b-kube-api-access-r7n2c\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862335 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/192c8b71-0927-4047-8fa6-a0b45fdbef7b-srv-cert\") pod \"olm-operator-6b444d44fb-sspwb\" (UID: \"192c8b71-0927-4047-8fa6-a0b45fdbef7b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862355 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdfkq\" (UniqueName: \"kubernetes.io/projected/4fc4931a-8c26-4f29-8798-d730e10e8d39-kube-api-access-sdfkq\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862371 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4zvs\" (UniqueName: \"kubernetes.io/projected/9de03415-ecf6-43a2-8968-eae338108559-kube-api-access-n4zvs\") pod \"service-ca-operator-777779d784-nslc6\" (UID: \"9de03415-ecf6-43a2-8968-eae338108559\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862395 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/eee9658e-289f-4ecf-8661-5b1bdf9c307c-metrics-tls\") pod \"dns-default-gw877\" (UID: \"eee9658e-289f-4ecf-8661-5b1bdf9c307c\") " pod="openshift-dns/dns-default-gw877" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862431 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cb535ff-3daa-418e-929e-99664eea736b-serving-cert\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862451 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3ae04268-6e5b-402f-b3c3-6b456df7e689-tmpfs\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862468 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dd6facda-112d-4218-9cec-3e24d3d2249d-proxy-tls\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862484 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862507 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-registration-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862528 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dd6facda-112d-4218-9cec-3e24d3d2249d-images\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862547 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fcc2c29a-2e97-457a-8f25-d96a4aee4ea5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pz4hd\" (UID: \"fcc2c29a-2e97-457a-8f25-d96a4aee4ea5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862564 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3ae04268-6e5b-402f-b3c3-6b456df7e689-apiservice-cert\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862612 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bfpg\" (UniqueName: \"kubernetes.io/projected/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-kube-api-access-6bfpg\") pod \"collect-profiles-29421645-tk9h9\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862637 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bspcd\" (UniqueName: \"kubernetes.io/projected/4e939827-530f-4e2e-9455-080369e5c76d-kube-api-access-bspcd\") pod \"catalog-operator-68c6474976-r7zd5\" (UID: \"4e939827-530f-4e2e-9455-080369e5c76d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862657 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df27667c-ecf7-4e33-9d34-c3a5b4088154-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-q4bd2\" (UID: \"df27667c-ecf7-4e33-9d34-c3a5b4088154\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862674 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/05f2d085-d179-42d9-bc77-7dce694be634-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zlqrk\" (UID: \"05f2d085-d179-42d9-bc77-7dce694be634\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862697 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ec10355-df85-4f79-9bd5-78ff480e25f4-etcd-service-ca\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862723 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g28g9\" (UniqueName: \"kubernetes.io/projected/8f80de45-61f5-4538-93d7-c6252e3c64bc-kube-api-access-g28g9\") pod \"control-plane-machine-set-operator-78cbb6b69f-7jpc4\" (UID: \"8f80de45-61f5-4538-93d7-c6252e3c64bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862750 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862770 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcr8d\" (UniqueName: \"kubernetes.io/projected/d30d8b55-366e-44df-967d-57e72d253fe6-kube-api-access-jcr8d\") pod \"machine-config-server-fvjws\" (UID: \"d30d8b55-366e-44df-967d-57e72d253fe6\") " pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862795 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/849c8fd6-0873-40d0-8c40-7bbc2365452d-trusted-ca\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862839 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/192c8b71-0927-4047-8fa6-a0b45fdbef7b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sspwb\" (UID: \"192c8b71-0927-4047-8fa6-a0b45fdbef7b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862865 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jch89\" (UniqueName: \"kubernetes.io/projected/afae1fb2-041c-433d-9e7b-e29a7e82ed31-kube-api-access-jch89\") pod \"package-server-manager-789f6589d5-cbfp9\" (UID: \"afae1fb2-041c-433d-9e7b-e29a7e82ed31\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862886 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mldc6\" (UniqueName: \"kubernetes.io/projected/ffebb621-9aa6-4104-a457-c180e6c93a28-kube-api-access-mldc6\") pod \"service-ca-9c57cc56f-qmwzp\" (UID: \"ffebb621-9aa6-4104-a457-c180e6c93a28\") " pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862906 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec10355-df85-4f79-9bd5-78ff480e25f4-serving-cert\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862915 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-config-volume\") pod \"collect-profiles-29421645-tk9h9\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862922 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-mountpoint-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862993 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b423076b-47af-4e4e-a085-d47074f6bb8f-cert\") pod \"ingress-canary-cmrfr\" (UID: \"b423076b-47af-4e4e-a085-d47074f6bb8f\") " pod="openshift-ingress-canary/ingress-canary-cmrfr" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.862992 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-mountpoint-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.863044 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eee9658e-289f-4ecf-8661-5b1bdf9c307c-config-volume\") pod \"dns-default-gw877\" (UID: \"eee9658e-289f-4ecf-8661-5b1bdf9c307c\") " pod="openshift-dns/dns-default-gw877" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.863081 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-config\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.863128 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dd6facda-112d-4218-9cec-3e24d3d2249d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.863156 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/afae1fb2-041c-433d-9e7b-e29a7e82ed31-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cbfp9\" (UID: \"afae1fb2-041c-433d-9e7b-e29a7e82ed31\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.865424 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-csi-data-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.866222 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cbebf2b4-7952-45d6-8c77-efc6438e05fd-stats-auth\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.866251 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhkx9\" (UniqueName: \"kubernetes.io/projected/9ada019a-2740-482e-864b-55762c330b26-kube-api-access-rhkx9\") pod \"oauth-openshift-558db77b4-6449l\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.866373 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4e939827-530f-4e2e-9455-080369e5c76d-profile-collector-cert\") pod \"catalog-operator-68c6474976-r7zd5\" (UID: \"4e939827-530f-4e2e-9455-080369e5c76d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.866469 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-client-ca\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.867147 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8jkvw\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.867434 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dd6facda-112d-4218-9cec-3e24d3d2249d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.868549 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df27667c-ecf7-4e33-9d34-c3a5b4088154-config\") pod \"kube-controller-manager-operator-78b949d7b-q4bd2\" (UID: \"df27667c-ecf7-4e33-9d34-c3a5b4088154\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.868870 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ec10355-df85-4f79-9bd5-78ff480e25f4-etcd-service-ca\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.869037 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d30d8b55-366e-44df-967d-57e72d253fe6-certs\") pod \"machine-config-server-fvjws\" (UID: \"d30d8b55-366e-44df-967d-57e72d253fe6\") " pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:10 crc kubenswrapper[4954]: E1209 16:59:10.869191 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.369176403 +0000 UTC m=+147.757350213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.870653 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/849c8fd6-0873-40d0-8c40-7bbc2365452d-trusted-ca\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.871494 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/192c8b71-0927-4047-8fa6-a0b45fdbef7b-srv-cert\") pod \"olm-operator-6b444d44fb-sspwb\" (UID: \"192c8b71-0927-4047-8fa6-a0b45fdbef7b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.872030 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-config\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.872341 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df27667c-ecf7-4e33-9d34-c3a5b4088154-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-q4bd2\" (UID: \"df27667c-ecf7-4e33-9d34-c3a5b4088154\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.873835 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/afae1fb2-041c-433d-9e7b-e29a7e82ed31-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cbfp9\" (UID: \"afae1fb2-041c-433d-9e7b-e29a7e82ed31\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.874607 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.876142 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dd6facda-112d-4218-9cec-3e24d3d2249d-images\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.877181 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-registration-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.877733 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f82b145c-6847-4a03-afff-2a513a79ecbe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-prclt\" (UID: \"f82b145c-6847-4a03-afff-2a513a79ecbe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.880066 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cb535ff-3daa-418e-929e-99664eea736b-serving-cert\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.880405 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3ae04268-6e5b-402f-b3c3-6b456df7e689-tmpfs\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.877738 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8jkvw\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.879395 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/eee9658e-289f-4ecf-8661-5b1bdf9c307c-metrics-tls\") pod \"dns-default-gw877\" (UID: \"eee9658e-289f-4ecf-8661-5b1bdf9c307c\") " pod="openshift-dns/dns-default-gw877" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.875562 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cbebf2b4-7952-45d6-8c77-efc6438e05fd-metrics-certs\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.884413 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-socket-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.885785 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec10355-df85-4f79-9bd5-78ff480e25f4-config\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.886408 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fcc2c29a-2e97-457a-8f25-d96a4aee4ea5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pz4hd\" (UID: \"fcc2c29a-2e97-457a-8f25-d96a4aee4ea5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.887518 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05f2d085-d179-42d9-bc77-7dce694be634-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zlqrk\" (UID: \"05f2d085-d179-42d9-bc77-7dce694be634\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.888187 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f82b145c-6847-4a03-afff-2a513a79ecbe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-prclt\" (UID: \"f82b145c-6847-4a03-afff-2a513a79ecbe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.888323 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4fc4931a-8c26-4f29-8798-d730e10e8d39-plugins-dir\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.877142 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eee9658e-289f-4ecf-8661-5b1bdf9c307c-config-volume\") pod \"dns-default-gw877\" (UID: \"eee9658e-289f-4ecf-8661-5b1bdf9c307c\") " pod="openshift-dns/dns-default-gw877" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.894143 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ffebb621-9aa6-4104-a457-c180e6c93a28-signing-key\") pod \"service-ca-9c57cc56f-qmwzp\" (UID: \"ffebb621-9aa6-4104-a457-c180e6c93a28\") " pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.894284 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.895092 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7ec10355-df85-4f79-9bd5-78ff480e25f4-etcd-ca\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.895813 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3ae04268-6e5b-402f-b3c3-6b456df7e689-webhook-cert\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.896191 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbebf2b4-7952-45d6-8c77-efc6438e05fd-service-ca-bundle\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.896231 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-secret-volume\") pod \"collect-profiles-29421645-tk9h9\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.897473 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d30d8b55-366e-44df-967d-57e72d253fe6-node-bootstrap-token\") pod \"machine-config-server-fvjws\" (UID: \"d30d8b55-366e-44df-967d-57e72d253fe6\") " pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.899246 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/192c8b71-0927-4047-8fa6-a0b45fdbef7b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sspwb\" (UID: \"192c8b71-0927-4047-8fa6-a0b45fdbef7b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.900626 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9de03415-ecf6-43a2-8968-eae338108559-config\") pod \"service-ca-operator-777779d784-nslc6\" (UID: \"9de03415-ecf6-43a2-8968-eae338108559\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.901283 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b423076b-47af-4e4e-a085-d47074f6bb8f-cert\") pod \"ingress-canary-cmrfr\" (UID: \"b423076b-47af-4e4e-a085-d47074f6bb8f\") " pod="openshift-ingress-canary/ingress-canary-cmrfr" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.902336 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec10355-df85-4f79-9bd5-78ff480e25f4-serving-cert\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.903533 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ffebb621-9aa6-4104-a457-c180e6c93a28-signing-cabundle\") pod \"service-ca-9c57cc56f-qmwzp\" (UID: \"ffebb621-9aa6-4104-a457-c180e6c93a28\") " pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.905188 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9de03415-ecf6-43a2-8968-eae338108559-serving-cert\") pod \"service-ca-operator-777779d784-nslc6\" (UID: \"9de03415-ecf6-43a2-8968-eae338108559\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.905312 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7ec10355-df85-4f79-9bd5-78ff480e25f4-etcd-client\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.905578 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cbebf2b4-7952-45d6-8c77-efc6438e05fd-default-certificate\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.905822 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05f2d085-d179-42d9-bc77-7dce694be634-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zlqrk\" (UID: \"05f2d085-d179-42d9-bc77-7dce694be634\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.905849 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/849c8fd6-0873-40d0-8c40-7bbc2365452d-metrics-tls\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.906249 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f80de45-61f5-4538-93d7-c6252e3c64bc-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7jpc4\" (UID: \"8f80de45-61f5-4538-93d7-c6252e3c64bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.906295 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4jct\" (UniqueName: \"kubernetes.io/projected/1df4a2c0-8c86-447c-aa38-cb9a436f84eb-kube-api-access-r4jct\") pod \"openshift-controller-manager-operator-756b6f6bc6-j9crx\" (UID: \"1df4a2c0-8c86-447c-aa38-cb9a436f84eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.906635 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dd6facda-112d-4218-9cec-3e24d3d2249d-proxy-tls\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.907143 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3ae04268-6e5b-402f-b3c3-6b456df7e689-apiservice-cert\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.913708 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4e939827-530f-4e2e-9455-080369e5c76d-srv-cert\") pod \"catalog-operator-68c6474976-r7zd5\" (UID: \"4e939827-530f-4e2e-9455-080369e5c76d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.926424 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-bound-sa-token\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.931071 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfrvc\" (UniqueName: \"kubernetes.io/projected/57dc805d-5ae9-4f59-80fb-0dd9e15a33e0-kube-api-access-nfrvc\") pod \"machine-config-controller-84d6567774-8gzxc\" (UID: \"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.953283 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp"] Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.955486 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv"] Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.964366 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:10 crc kubenswrapper[4954]: E1209 16:59:10.964902 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.464884493 +0000 UTC m=+147.853058313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.972045 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd7jd\" (UniqueName: \"kubernetes.io/projected/dd6facda-112d-4218-9cec-3e24d3d2249d-kube-api-access-dd7jd\") pod \"machine-config-operator-74547568cd-chrbl\" (UID: \"dd6facda-112d-4218-9cec-3e24d3d2249d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.988806 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bll4m\" (UniqueName: \"kubernetes.io/projected/849c8fd6-0873-40d0-8c40-7bbc2365452d-kube-api-access-bll4m\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.989751 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6gmvm"] Dec 09 16:59:10 crc kubenswrapper[4954]: I1209 16:59:10.990866 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-96v5v"] Dec 09 16:59:10 crc kubenswrapper[4954]: W1209 16:59:10.999686 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda759536e_62a7_4410_b1a3_0ea059cd63f4.slice/crio-0fefe3d39fb533e7770027ca72fb5923fcaf6f21b31da5682c6ca4199bef6c11 WatchSource:0}: Error finding container 0fefe3d39fb533e7770027ca72fb5923fcaf6f21b31da5682c6ca4199bef6c11: Status 404 returned error can't find the container with id 0fefe3d39fb533e7770027ca72fb5923fcaf6f21b31da5682c6ca4199bef6c11 Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.002004 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.007139 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.007831 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7n2c\" (UniqueName: \"kubernetes.io/projected/1cb535ff-3daa-418e-929e-99664eea736b-kube-api-access-r7n2c\") pod \"controller-manager-879f6c89f-mnc4f\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:11 crc kubenswrapper[4954]: W1209 16:59:11.007940 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a770208_d33d_48d7_b70c_be193682fb72.slice/crio-53f691fbd0badf79c0d8b31af8226fb73c095d16cfcd33b5b2df0d0762f0f6f3 WatchSource:0}: Error finding container 53f691fbd0badf79c0d8b31af8226fb73c095d16cfcd33b5b2df0d0762f0f6f3: Status 404 returned error can't find the container with id 53f691fbd0badf79c0d8b31af8226fb73c095d16cfcd33b5b2df0d0762f0f6f3 Dec 09 16:59:11 crc kubenswrapper[4954]: W1209 16:59:11.008167 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc84a7cef_2031_4477_adbb_0e5f97cf7726.slice/crio-20c292b1b9a4da43d106f41163455cda7a5d7adfe9e6c704a2db35b593cd32d9 WatchSource:0}: Error finding container 20c292b1b9a4da43d106f41163455cda7a5d7adfe9e6c704a2db35b593cd32d9: Status 404 returned error can't find the container with id 20c292b1b9a4da43d106f41163455cda7a5d7adfe9e6c704a2db35b593cd32d9 Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.009940 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-v4pm7"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.012477 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.021959 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.023636 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bfpg\" (UniqueName: \"kubernetes.io/projected/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-kube-api-access-6bfpg\") pod \"collect-profiles-29421645-tk9h9\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:11 crc kubenswrapper[4954]: W1209 16:59:11.027237 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad6309f7_1627_4f65_9169_181e7a046adc.slice/crio-01d47e53af5bcea078a1d082bf355a6950cb56d9b50c656636a6b91c96c781c4 WatchSource:0}: Error finding container 01d47e53af5bcea078a1d082bf355a6950cb56d9b50c656636a6b91c96c781c4: Status 404 returned error can't find the container with id 01d47e53af5bcea078a1d082bf355a6950cb56d9b50c656636a6b91c96c781c4 Dec 09 16:59:11 crc kubenswrapper[4954]: W1209 16:59:11.036695 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bb5ad5d_2235_47f8_8100_8b79d0d44fb0.slice/crio-317339b18d0b617eab6ed4f455e51414fe7e42cf6a08bd12f2606b411b42df6b WatchSource:0}: Error finding container 317339b18d0b617eab6ed4f455e51414fe7e42cf6a08bd12f2606b411b42df6b: Status 404 returned error can't find the container with id 317339b18d0b617eab6ed4f455e51414fe7e42cf6a08bd12f2606b411b42df6b Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.066384 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.066891 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.56687099 +0000 UTC m=+147.955044870 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.067798 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdfkq\" (UniqueName: \"kubernetes.io/projected/4fc4931a-8c26-4f29-8798-d730e10e8d39-kube-api-access-sdfkq\") pod \"csi-hostpathplugin-b4gqt\" (UID: \"4fc4931a-8c26-4f29-8798-d730e10e8d39\") " pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.072424 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4zvs\" (UniqueName: \"kubernetes.io/projected/9de03415-ecf6-43a2-8968-eae338108559-kube-api-access-n4zvs\") pod \"service-ca-operator-777779d784-nslc6\" (UID: \"9de03415-ecf6-43a2-8968-eae338108559\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.089972 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.094623 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-v86j2"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.099495 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bspcd\" (UniqueName: \"kubernetes.io/projected/4e939827-530f-4e2e-9455-080369e5c76d-kube-api-access-bspcd\") pod \"catalog-operator-68c6474976-r7zd5\" (UID: \"4e939827-530f-4e2e-9455-080369e5c76d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.103107 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.104435 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.106945 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.111434 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df27667c-ecf7-4e33-9d34-c3a5b4088154-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-q4bd2\" (UID: \"df27667c-ecf7-4e33-9d34-c3a5b4088154\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.122649 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr"] Dec 09 16:59:11 crc kubenswrapper[4954]: W1209 16:59:11.123007 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73c4eb3e_d7b2_48fd_a07f_6b73dc18ccfe.slice/crio-0c92f2066a26006a582a62e875cd7e46514a610ce0d7d1123df14fe08665f2ab WatchSource:0}: Error finding container 0c92f2066a26006a582a62e875cd7e46514a610ce0d7d1123df14fe08665f2ab: Status 404 returned error can't find the container with id 0c92f2066a26006a582a62e875cd7e46514a610ce0d7d1123df14fe08665f2ab Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.135874 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/05f2d085-d179-42d9-bc77-7dce694be634-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zlqrk\" (UID: \"05f2d085-d179-42d9-bc77-7dce694be634\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.139644 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.139924 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.146220 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.147496 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g28g9\" (UniqueName: \"kubernetes.io/projected/8f80de45-61f5-4538-93d7-c6252e3c64bc-kube-api-access-g28g9\") pod \"control-plane-machine-set-operator-78cbb6b69f-7jpc4\" (UID: \"8f80de45-61f5-4538-93d7-c6252e3c64bc\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" Dec 09 16:59:11 crc kubenswrapper[4954]: W1209 16:59:11.160767 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a1f21d3_3aee_45dc_baa6_2929b5f6e150.slice/crio-5b142e1f0dbbfa2940c088d060538e97aad1d5512d93ffe539890a5c957c1575 WatchSource:0}: Error finding container 5b142e1f0dbbfa2940c088d060538e97aad1d5512d93ffe539890a5c957c1575: Status 404 returned error can't find the container with id 5b142e1f0dbbfa2940c088d060538e97aad1d5512d93ffe539890a5c957c1575 Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.168967 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.169243 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.169289 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.169335 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.169463 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.171452 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.173241 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcr8d\" (UniqueName: \"kubernetes.io/projected/d30d8b55-366e-44df-967d-57e72d253fe6-kube-api-access-jcr8d\") pod \"machine-config-server-fvjws\" (UID: \"d30d8b55-366e-44df-967d-57e72d253fe6\") " pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.173368 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.673346954 +0000 UTC m=+148.061520844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.175305 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.177197 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.178151 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.195139 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.197171 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mldc6\" (UniqueName: \"kubernetes.io/projected/ffebb621-9aa6-4104-a457-c180e6c93a28-kube-api-access-mldc6\") pod \"service-ca-9c57cc56f-qmwzp\" (UID: \"ffebb621-9aa6-4104-a457-c180e6c93a28\") " pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.203181 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jch89\" (UniqueName: \"kubernetes.io/projected/afae1fb2-041c-433d-9e7b-e29a7e82ed31-kube-api-access-jch89\") pod \"package-server-manager-789f6589d5-cbfp9\" (UID: \"afae1fb2-041c-433d-9e7b-e29a7e82ed31\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.228053 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2pvl\" (UniqueName: \"kubernetes.io/projected/cbebf2b4-7952-45d6-8c77-efc6438e05fd-kube-api-access-g2pvl\") pod \"router-default-5444994796-cj74l\" (UID: \"cbebf2b4-7952-45d6-8c77-efc6438e05fd\") " pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.237853 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fvjws" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.246606 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr8c9\" (UniqueName: \"kubernetes.io/projected/b423076b-47af-4e4e-a085-d47074f6bb8f-kube-api-access-hr8c9\") pod \"ingress-canary-cmrfr\" (UID: \"b423076b-47af-4e4e-a085-d47074f6bb8f\") " pod="openshift-ingress-canary/ingress-canary-cmrfr" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.264683 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.270445 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.270878 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.770863085 +0000 UTC m=+148.159036905 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.272906 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2qm8\" (UniqueName: \"kubernetes.io/projected/fcc2c29a-2e97-457a-8f25-d96a4aee4ea5-kube-api-access-s2qm8\") pod \"multus-admission-controller-857f4d67dd-pz4hd\" (UID: \"fcc2c29a-2e97-457a-8f25-d96a4aee4ea5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.293440 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x52dh\" (UniqueName: \"kubernetes.io/projected/192c8b71-0927-4047-8fa6-a0b45fdbef7b-kube-api-access-x52dh\") pod \"olm-operator-6b444d44fb-sspwb\" (UID: \"192c8b71-0927-4047-8fa6-a0b45fdbef7b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.309126 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4brz\" (UniqueName: \"kubernetes.io/projected/7ec10355-df85-4f79-9bd5-78ff480e25f4-kube-api-access-q4brz\") pod \"etcd-operator-b45778765-mhkmt\" (UID: \"7ec10355-df85-4f79-9bd5-78ff480e25f4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.326532 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwphl\" (UniqueName: \"kubernetes.io/projected/f82b145c-6847-4a03-afff-2a513a79ecbe-kube-api-access-kwphl\") pod \"kube-storage-version-migrator-operator-b67b599dd-prclt\" (UID: \"f82b145c-6847-4a03-afff-2a513a79ecbe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.337130 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.344903 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.350878 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.354964 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/849c8fd6-0873-40d0-8c40-7bbc2365452d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lz9lt\" (UID: \"849c8fd6-0873-40d0-8c40-7bbc2365452d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.364504 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.371930 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.373183 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.373275 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.873251875 +0000 UTC m=+148.261425745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.377750 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.378165 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.878139624 +0000 UTC m=+148.266313444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.379316 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.381293 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8rxv\" (UniqueName: \"kubernetes.io/projected/eee9658e-289f-4ecf-8661-5b1bdf9c307c-kube-api-access-n8rxv\") pod \"dns-default-gw877\" (UID: \"eee9658e-289f-4ecf-8661-5b1bdf9c307c\") " pod="openshift-dns/dns-default-gw877" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.384401 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.393038 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.393837 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nq2tq"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.397177 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnqqf\" (UniqueName: \"kubernetes.io/projected/454e49f6-b054-4901-bd18-461d7e9cd89e-kube-api-access-rnqqf\") pod \"marketplace-operator-79b997595-8jkvw\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.401280 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.416283 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.416566 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrwq6\" (UniqueName: \"kubernetes.io/projected/3ae04268-6e5b-402f-b3c3-6b456df7e689-kube-api-access-mrwq6\") pod \"packageserver-d55dfcdfc-pp8mx\" (UID: \"3ae04268-6e5b-402f-b3c3-6b456df7e689\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.427215 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.431880 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.454601 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.464739 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.485000 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.485773 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.486890 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.986865921 +0000 UTC m=+148.375039741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.490205 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.492685 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:11.992665926 +0000 UTC m=+148.380839746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.498185 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.506158 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.517083 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cmrfr" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.529107 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gw877" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.546909 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.594744 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.595083 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.095067456 +0000 UTC m=+148.483241276 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.644921 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.696574 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.697044 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.197031803 +0000 UTC m=+148.585205623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: W1209 16:59:11.719956 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1df4a2c0_8c86_447c_aa38_cb9a436f84eb.slice/crio-b8e22f89cf5f8da77daa91458c993be11663bde1fe5482dbebf447ed5de4f2e3 WatchSource:0}: Error finding container b8e22f89cf5f8da77daa91458c993be11663bde1fe5482dbebf447ed5de4f2e3: Status 404 returned error can't find the container with id b8e22f89cf5f8da77daa91458c993be11663bde1fe5482dbebf447ed5de4f2e3 Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.733061 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mnc4f"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.758629 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-v6ftv"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.800556 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.800927 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.300910212 +0000 UTC m=+148.689084032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.861310 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" event={"ID":"ad6309f7-1627-4f65-9169-181e7a046adc","Type":"ContainerStarted","Data":"01d47e53af5bcea078a1d082bf355a6950cb56d9b50c656636a6b91c96c781c4"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.862443 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-v86j2" event={"ID":"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe","Type":"ContainerStarted","Data":"f6f852a5597e6462c638004bc4360e08aa12da5379b44c63d2055d6f7feaa427"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.862464 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-v86j2" event={"ID":"73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe","Type":"ContainerStarted","Data":"0c92f2066a26006a582a62e875cd7e46514a610ce0d7d1123df14fe08665f2ab"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.862733 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.879399 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" event={"ID":"cd86b184-0ad7-416c-a326-70dad6e39296","Type":"ContainerStarted","Data":"bf2f389532e5f87abf36cc18107c69b37c502792b0076a8356e7120fac4fb8c6"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.886950 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" event={"ID":"6a770208-d33d-48d7-b70c-be193682fb72","Type":"ContainerStarted","Data":"0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.887000 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" event={"ID":"6a770208-d33d-48d7-b70c-be193682fb72","Type":"ContainerStarted","Data":"53f691fbd0badf79c0d8b31af8226fb73c095d16cfcd33b5b2df0d0762f0f6f3"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.887378 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.893045 4954 patch_prober.go:28] interesting pod/console-operator-58897d9998-v86j2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.893121 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-v86j2" podUID="73c4eb3e-d7b2-48fd-a07f-6b73dc18ccfe" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.902767 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:11 crc kubenswrapper[4954]: E1209 16:59:11.904249 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.404230957 +0000 UTC m=+148.792404777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.906555 4954 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-5cftp container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.906651 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" podUID="6a770208-d33d-48d7-b70c-be193682fb72" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.909713 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.936729 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" event={"ID":"db03f79f-fcea-4a92-b021-bac0b6c1cebc","Type":"ContainerStarted","Data":"13b55f0adc6d5e10de9da882e17d67de62e050dd7f7b7bf9c3e233ade29b4e49"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.936789 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" event={"ID":"db03f79f-fcea-4a92-b021-bac0b6c1cebc","Type":"ContainerStarted","Data":"9a356c3c20a66a249f99f7d066809b2831df6fad82c974fb860cdce6cbae7465"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.951960 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" event={"ID":"c84a7cef-2031-4477-adbb-0e5f97cf7726","Type":"ContainerStarted","Data":"9f8fb873f8223a918ab34a23f18afb393ba3f68ec61f9d5bd724d0e78c25954f"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.952033 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" event={"ID":"c84a7cef-2031-4477-adbb-0e5f97cf7726","Type":"ContainerStarted","Data":"20c292b1b9a4da43d106f41163455cda7a5d7adfe9e6c704a2db35b593cd32d9"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.957942 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" event={"ID":"1cb535ff-3daa-418e-929e-99664eea736b","Type":"ContainerStarted","Data":"37a701a282458902b44da1488e10efb91809ae98d999e159ae4431c94bcde645"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.964181 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" event={"ID":"1df4a2c0-8c86-447c-aa38-cb9a436f84eb","Type":"ContainerStarted","Data":"b8e22f89cf5f8da77daa91458c993be11663bde1fe5482dbebf447ed5de4f2e3"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.967776 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6449l"] Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.971574 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" event={"ID":"a1de74a6-26b3-46df-b9bf-717509ae4150","Type":"ContainerStarted","Data":"79e732c6e38bfc1a29bd69f0a4bb408cf88686c89f197eede138733d752f350c"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.971663 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" event={"ID":"a1de74a6-26b3-46df-b9bf-717509ae4150","Type":"ContainerStarted","Data":"22f2cb07057dead5e25fa300bbb36613fa926737c5bbca66b31dc92999093050"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.991352 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" event={"ID":"a759536e-62a7-4410-b1a3-0ea059cd63f4","Type":"ContainerStarted","Data":"087a67d1c3618b402a8b8a03eb50b014b7c26cd928f416738119f3fd358f36bb"} Dec 09 16:59:11 crc kubenswrapper[4954]: I1209 16:59:11.991422 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" event={"ID":"a759536e-62a7-4410-b1a3-0ea059cd63f4","Type":"ContainerStarted","Data":"0fefe3d39fb533e7770027ca72fb5923fcaf6f21b31da5682c6ca4199bef6c11"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.005352 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.005566 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.505539061 +0000 UTC m=+148.893712881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.005775 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" event={"ID":"8ffdbd23-8e68-4bf3-bb37-2817fa11f18b","Type":"ContainerStarted","Data":"2b05cb21f99a4654764de9cdfa21a860fb875051c466472bedb30d14efd76fb4"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.005845 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.007857 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nslc6"] Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.007938 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.507911205 +0000 UTC m=+148.896085025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.023546 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fvjws" event={"ID":"d30d8b55-366e-44df-967d-57e72d253fe6","Type":"ContainerStarted","Data":"15c88f5ae7a220294740d3722546ff2201a0eda5660040b05623155bfc505e85"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.037625 4954 generic.go:334] "Generic (PLEG): container finished" podID="ff5d80e6-ad76-49c5-a350-c7cd9e8528ae" containerID="122de7e3f66e0c440860b7301092083bfce3ca440150c508f09adde337276313" exitCode=0 Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.037782 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" event={"ID":"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae","Type":"ContainerDied","Data":"122de7e3f66e0c440860b7301092083bfce3ca440150c508f09adde337276313"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.047057 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl"] Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.052940 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6gmvm" event={"ID":"dc11e3bb-97da-485b-ad28-34b0116ed220","Type":"ContainerStarted","Data":"a49768838bf2298ccc0abf590b58a6ff051c216609e208f0dfdeead43339ee1a"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.052983 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6gmvm" event={"ID":"dc11e3bb-97da-485b-ad28-34b0116ed220","Type":"ContainerStarted","Data":"01720932d12c84593be4162fdd0effbf1bda558540bcac9d68b0abefbcd448eb"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.053359 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.055060 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.055101 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.065905 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" event={"ID":"82cdcdae-1e7b-4591-80f7-b297bef6d91f","Type":"ContainerStarted","Data":"0d1934bf2006e2061deb51c811ad4151d25186d14a57de44dc0c0fe458b09d6d"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.065957 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" event={"ID":"82cdcdae-1e7b-4591-80f7-b297bef6d91f","Type":"ContainerStarted","Data":"31ccaf1a5157662abe3746f49db6275a02bef2eedbea24a99a308b5f6d9c1127"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.071979 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr" event={"ID":"4a1f21d3-3aee-45dc-baa6-2929b5f6e150","Type":"ContainerStarted","Data":"3d84bc2d6fd2904c76a2a1edc29f7652aa6195bde41d4d2ac1e4e78bf63ffec9"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.072020 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr" event={"ID":"4a1f21d3-3aee-45dc-baa6-2929b5f6e150","Type":"ContainerStarted","Data":"5b142e1f0dbbfa2940c088d060538e97aad1d5512d93ffe539890a5c957c1575"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.072912 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" event={"ID":"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0","Type":"ContainerStarted","Data":"42046cf763e921d566bd2364a7ba3732f01a8a1b9e2ef2f30e46cb544344bb1c"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.081507 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-v4pm7" event={"ID":"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0","Type":"ContainerStarted","Data":"244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.081543 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-v4pm7" event={"ID":"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0","Type":"ContainerStarted","Data":"317339b18d0b617eab6ed4f455e51414fe7e42cf6a08bd12f2606b411b42df6b"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.086738 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" event={"ID":"1623d36d-33ed-4c91-9069-aba484b54a83","Type":"ContainerStarted","Data":"20a29ccf289ffb19634f2e94d2fabae8c8a7d70ce39ee72aca90c65af0ee9938"} Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.107523 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.109342 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.609319099 +0000 UTC m=+148.997492919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.209044 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.209782 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.709767187 +0000 UTC m=+149.097941007 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: W1209 16:59:12.230441 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ada019a_2740_482e_864b_55762c330b26.slice/crio-5537b743e782b6973efa1383bf6201177d90e5a128bcb21306a04919ec7ed18f WatchSource:0}: Error finding container 5537b743e782b6973efa1383bf6201177d90e5a128bcb21306a04919ec7ed18f: Status 404 returned error can't find the container with id 5537b743e782b6973efa1383bf6201177d90e5a128bcb21306a04919ec7ed18f Dec 09 16:59:12 crc kubenswrapper[4954]: W1209 16:59:12.291245 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9de03415_ecf6_43a2_8968_eae338108559.slice/crio-3c4743d49ac082e68b037bb117d15195f9817e0be16a405c2929a421075dc76c WatchSource:0}: Error finding container 3c4743d49ac082e68b037bb117d15195f9817e0be16a405c2929a421075dc76c: Status 404 returned error can't find the container with id 3c4743d49ac082e68b037bb117d15195f9817e0be16a405c2929a421075dc76c Dec 09 16:59:12 crc kubenswrapper[4954]: W1209 16:59:12.300585 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd6facda_112d_4218_9cec_3e24d3d2249d.slice/crio-e90450b380cfbd1596a2dc26c45687f0b71fdf95717f1659ac3d7b522db1cebd WatchSource:0}: Error finding container e90450b380cfbd1596a2dc26c45687f0b71fdf95717f1659ac3d7b522db1cebd: Status 404 returned error can't find the container with id e90450b380cfbd1596a2dc26c45687f0b71fdf95717f1659ac3d7b522db1cebd Dec 09 16:59:12 crc kubenswrapper[4954]: W1209 16:59:12.306777 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-5e6500e233dd8ae7c55b3a8176b9b29e1ca00ae463229b09e39af50f422585eb WatchSource:0}: Error finding container 5e6500e233dd8ae7c55b3a8176b9b29e1ca00ae463229b09e39af50f422585eb: Status 404 returned error can't find the container with id 5e6500e233dd8ae7c55b3a8176b9b29e1ca00ae463229b09e39af50f422585eb Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.314790 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.314976 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.814957834 +0000 UTC m=+149.203131654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.315072 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.315425 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.815416756 +0000 UTC m=+149.203590576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.394480 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2"] Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.414211 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.416849 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.417021 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.916972111 +0000 UTC m=+149.305145941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.417171 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.417579 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:12.917568795 +0000 UTC m=+149.305742615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.518222 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.518916 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.018873198 +0000 UTC m=+149.407047018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.519194 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.521071 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.021052571 +0000 UTC m=+149.409226391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.621351 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.621698 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.12168323 +0000 UTC m=+149.509857050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.705097 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" podStartSLOduration=129.705076357 podStartE2EDuration="2m9.705076357s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:12.704214752 +0000 UTC m=+149.092388592" watchObservedRunningTime="2025-12-09 16:59:12.705076357 +0000 UTC m=+149.093250177" Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.724587 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.724942 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.224926115 +0000 UTC m=+149.613099935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.826099 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.826538 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.326510331 +0000 UTC m=+149.714684151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.826786 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.827170 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.327155775 +0000 UTC m=+149.715329595 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.892858 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w2spv" podStartSLOduration=130.892837305 podStartE2EDuration="2m10.892837305s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:12.841724941 +0000 UTC m=+149.229898771" watchObservedRunningTime="2025-12-09 16:59:12.892837305 +0000 UTC m=+149.281011125" Dec 09 16:59:12 crc kubenswrapper[4954]: I1209 16:59:12.934349 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:12 crc kubenswrapper[4954]: E1209 16:59:12.934775 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.434759785 +0000 UTC m=+149.822933605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.009166 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nzwb5" podStartSLOduration=131.009132928 podStartE2EDuration="2m11.009132928s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:13.008509055 +0000 UTC m=+149.396682875" watchObservedRunningTime="2025-12-09 16:59:13.009132928 +0000 UTC m=+149.397306748" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.045344 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:13 crc kubenswrapper[4954]: E1209 16:59:13.045919 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.545904857 +0000 UTC m=+149.934078677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.057052 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6gmvm" podStartSLOduration=131.057009693 podStartE2EDuration="2m11.057009693s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:13.044017056 +0000 UTC m=+149.432190876" watchObservedRunningTime="2025-12-09 16:59:13.057009693 +0000 UTC m=+149.445183513" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.096430 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-h7bjt" podStartSLOduration=130.096405567 podStartE2EDuration="2m10.096405567s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:13.074433037 +0000 UTC m=+149.462606857" watchObservedRunningTime="2025-12-09 16:59:13.096405567 +0000 UTC m=+149.484579387" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.146132 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:13 crc kubenswrapper[4954]: E1209 16:59:13.146470 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.646452885 +0000 UTC m=+150.034626705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.168188 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" event={"ID":"df27667c-ecf7-4e33-9d34-c3a5b4088154","Type":"ContainerStarted","Data":"dd0f320a988ee15fd6fd00d8a97eb54dc8451af32a769a0a3933760ecd2f6c7c"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.183812 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" event={"ID":"cd86b184-0ad7-416c-a326-70dad6e39296","Type":"ContainerStarted","Data":"9342192432bbb2b1309ed8d90fe6a7b686950afd154777e578e3f0c5fcef870d"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.196848 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-v4pm7" podStartSLOduration=131.196827315 podStartE2EDuration="2m11.196827315s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:13.152975125 +0000 UTC m=+149.541148945" watchObservedRunningTime="2025-12-09 16:59:13.196827315 +0000 UTC m=+149.585001135" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.200939 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5e6500e233dd8ae7c55b3a8176b9b29e1ca00ae463229b09e39af50f422585eb"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.215913 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" event={"ID":"9de03415-ecf6-43a2-8968-eae338108559","Type":"ContainerStarted","Data":"3c4743d49ac082e68b037bb117d15195f9817e0be16a405c2929a421075dc76c"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.223387 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr" event={"ID":"4a1f21d3-3aee-45dc-baa6-2929b5f6e150","Type":"ContainerStarted","Data":"99b4811d816e6e537f375cdd293817da9b3127b0e69854b08ba47f356723a38a"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.233476 4954 generic.go:334] "Generic (PLEG): container finished" podID="db03f79f-fcea-4a92-b021-bac0b6c1cebc" containerID="13b55f0adc6d5e10de9da882e17d67de62e050dd7f7b7bf9c3e233ade29b4e49" exitCode=0 Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.233634 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" event={"ID":"db03f79f-fcea-4a92-b021-bac0b6c1cebc","Type":"ContainerDied","Data":"13b55f0adc6d5e10de9da882e17d67de62e050dd7f7b7bf9c3e233ade29b4e49"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.235248 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" event={"ID":"e43b0681-33e3-4eca-a737-ed3c9c2fe994","Type":"ContainerStarted","Data":"1294c71e167db0f0b0c0c682084b455ad273292f5d8f9b4ce2a381cebdf5cf69"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.251476 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.251539 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fvjws" event={"ID":"d30d8b55-366e-44df-967d-57e72d253fe6","Type":"ContainerStarted","Data":"3a89970691894c7d602d4b9b885021dfa755b0dae580ae48b2ab2e7427cfe080"} Dec 09 16:59:13 crc kubenswrapper[4954]: E1209 16:59:13.252324 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.752307417 +0000 UTC m=+150.140481237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.268475 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" event={"ID":"9ada019a-2740-482e-864b-55762c330b26","Type":"ContainerStarted","Data":"5537b743e782b6973efa1383bf6201177d90e5a128bcb21306a04919ec7ed18f"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.269621 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" event={"ID":"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81","Type":"ContainerStarted","Data":"d555a1e11e57675734c870b16bb4afbee52903dc26ab5e8a1feaa1df43e0d91c"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.271717 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" event={"ID":"dd6facda-112d-4218-9cec-3e24d3d2249d","Type":"ContainerStarted","Data":"e90450b380cfbd1596a2dc26c45687f0b71fdf95717f1659ac3d7b522db1cebd"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.279532 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" event={"ID":"1623d36d-33ed-4c91-9069-aba484b54a83","Type":"ContainerStarted","Data":"89b95e65a149bf68ce5d16accc32c004c6d4e02793e82c0deadd7a3379d20754"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.299423 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hfd9b" podStartSLOduration=131.299404737 podStartE2EDuration="2m11.299404737s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:13.297626636 +0000 UTC m=+149.685800466" watchObservedRunningTime="2025-12-09 16:59:13.299404737 +0000 UTC m=+149.687578557" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.305146 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" event={"ID":"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0","Type":"ContainerStarted","Data":"0ae75f0a174ab5d474c8b610ab514f1590c831f6b5dd1ae163ba2c20fc569642"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.319473 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" event={"ID":"ad6309f7-1627-4f65-9169-181e7a046adc","Type":"ContainerStarted","Data":"a68ba2a180128f8406d2769f4fb53a3f9421b7796ebbbcc6f5b30d7c0cb5b14b"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.322478 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cj74l" event={"ID":"cbebf2b4-7952-45d6-8c77-efc6438e05fd","Type":"ContainerStarted","Data":"3bd2daa8b92c3c126132d5054e7c87cfb2d7366ddbe73181dab9a2095087f3cc"} Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.324430 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.324465 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.354477 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:13 crc kubenswrapper[4954]: E1209 16:59:13.355203 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.855187299 +0000 UTC m=+150.243361119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.387917 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-v86j2" podStartSLOduration=131.387904024 podStartE2EDuration="2m11.387904024s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:13.386997479 +0000 UTC m=+149.775171309" watchObservedRunningTime="2025-12-09 16:59:13.387904024 +0000 UTC m=+149.776077844" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.417613 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-nq2tq" podStartSLOduration=131.417580251 podStartE2EDuration="2m11.417580251s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:13.416342453 +0000 UTC m=+149.804516273" watchObservedRunningTime="2025-12-09 16:59:13.417580251 +0000 UTC m=+149.805754071" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.443189 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gcjjz" podStartSLOduration=131.443172503 podStartE2EDuration="2m11.443172503s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:13.441211661 +0000 UTC m=+149.829385491" watchObservedRunningTime="2025-12-09 16:59:13.443172503 +0000 UTC m=+149.831346323" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.456721 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:13 crc kubenswrapper[4954]: E1209 16:59:13.463239 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:13.963225402 +0000 UTC m=+150.351399222 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.463778 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-fvjws" podStartSLOduration=5.463764255 podStartE2EDuration="5.463764255s" podCreationTimestamp="2025-12-09 16:59:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:13.463290362 +0000 UTC m=+149.851464182" watchObservedRunningTime="2025-12-09 16:59:13.463764255 +0000 UTC m=+149.851938075" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.557297 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:13 crc kubenswrapper[4954]: E1209 16:59:13.557796 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:14.057776446 +0000 UTC m=+150.445950266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.580477 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.602561 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5dhxr" podStartSLOduration=131.602542463 podStartE2EDuration="2m11.602542463s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:13.489048326 +0000 UTC m=+149.877222146" watchObservedRunningTime="2025-12-09 16:59:13.602542463 +0000 UTC m=+149.990716293" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.659396 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:13 crc kubenswrapper[4954]: E1209 16:59:13.659751 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:14.159739923 +0000 UTC m=+150.547913733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.674923 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-v86j2" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.747242 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mhkmt"] Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.756675 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.756730 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.760048 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt"] Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.760491 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:13 crc kubenswrapper[4954]: E1209 16:59:13.760909 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:14.260889685 +0000 UTC m=+150.649063505 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.764105 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9"] Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.817880 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gw877"] Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.847511 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-b4gqt"] Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.851117 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qmwzp"] Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.862044 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:13 crc kubenswrapper[4954]: E1209 16:59:13.862402 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:14.36238972 +0000 UTC m=+150.750563540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.911502 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cmrfr"] Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.911559 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb"] Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.948627 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx"] Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.956199 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4"] Dec 09 16:59:13 crc kubenswrapper[4954]: W1209 16:59:13.959714 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-951511b8d5444911e3cf4125d03a3ce4c6c674515afcd4ef8683d56f0fe4fa95 WatchSource:0}: Error finding container 951511b8d5444911e3cf4125d03a3ce4c6c674515afcd4ef8683d56f0fe4fa95: Status 404 returned error can't find the container with id 951511b8d5444911e3cf4125d03a3ce4c6c674515afcd4ef8683d56f0fe4fa95 Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.963427 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:13 crc kubenswrapper[4954]: E1209 16:59:13.963798 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:14.463768764 +0000 UTC m=+150.851942584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.987216 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5"] Dec 09 16:59:13 crc kubenswrapper[4954]: I1209 16:59:13.996686 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8jkvw"] Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:13.999639 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk"] Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.001459 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt"] Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.003762 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pz4hd"] Dec 09 16:59:14 crc kubenswrapper[4954]: W1209 16:59:14.040347 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ae04268_6e5b_402f_b3c3_6b456df7e689.slice/crio-3812d624dcf788489f194699cf937047d8d590ddc598e6d3b6dfe5f2a921d535 WatchSource:0}: Error finding container 3812d624dcf788489f194699cf937047d8d590ddc598e6d3b6dfe5f2a921d535: Status 404 returned error can't find the container with id 3812d624dcf788489f194699cf937047d8d590ddc598e6d3b6dfe5f2a921d535 Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.066070 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:14 crc kubenswrapper[4954]: E1209 16:59:14.066500 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:14.566484745 +0000 UTC m=+150.954658565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.167086 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:14 crc kubenswrapper[4954]: E1209 16:59:14.167642 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:14.667620298 +0000 UTC m=+151.055794118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.272513 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:14 crc kubenswrapper[4954]: E1209 16:59:14.273180 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:14.773165976 +0000 UTC m=+151.161339796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.373562 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:14 crc kubenswrapper[4954]: E1209 16:59:14.373845 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:14.873830626 +0000 UTC m=+151.262004446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.440143 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" event={"ID":"1cb535ff-3daa-418e-929e-99664eea736b","Type":"ContainerStarted","Data":"121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.441660 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.448370 4954 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-mnc4f container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.448423 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" podUID="1cb535ff-3daa-418e-929e-99664eea736b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.449472 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"021d219139165de77581e54a45400c33c1824cfc7d5f6aa371aaf8b6a5752630"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.462332 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f36cf6588c38553f1c61438a83825989b48e901990d7b42e50c035f5860c5d5b"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.462400 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"951511b8d5444911e3cf4125d03a3ce4c6c674515afcd4ef8683d56f0fe4fa95"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.462987 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.476711 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:14 crc kubenswrapper[4954]: E1209 16:59:14.477748 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:14.977731175 +0000 UTC m=+151.365904995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.501407 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" event={"ID":"cd86b184-0ad7-416c-a326-70dad6e39296","Type":"ContainerStarted","Data":"814ec85b167c7223c018fabbb9d1668680af370e7deae42027ca53046c3a87f1"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.520957 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" event={"ID":"fcc2c29a-2e97-457a-8f25-d96a4aee4ea5","Type":"ContainerStarted","Data":"62e1ef5517792cfbb72dfec200d360f186e55de3da894eaf845032a9a9b3012e"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.533384 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" event={"ID":"7ec10355-df85-4f79-9bd5-78ff480e25f4","Type":"ContainerStarted","Data":"15fe4057de6b5962f4f6bed9896c8bb8d01be2fd882caac18258997c53ef6d9e"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.578447 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:14 crc kubenswrapper[4954]: E1209 16:59:14.580069 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:15.080027353 +0000 UTC m=+151.468201173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.601266 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" event={"ID":"849c8fd6-0873-40d0-8c40-7bbc2365452d","Type":"ContainerStarted","Data":"3f9a46a8510d870ee62813af21195f8f7d6454bbfcf60719983917659d247f15"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.611894 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" event={"ID":"db03f79f-fcea-4a92-b021-bac0b6c1cebc","Type":"ContainerStarted","Data":"3ed0c42efbcbf15af2ade525c4bba1cf9e7529bfec957bd3c670889a5f8aca66"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.613821 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.682733 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:14 crc kubenswrapper[4954]: E1209 16:59:14.683320 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:15.183300279 +0000 UTC m=+151.571474099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.721178 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" event={"ID":"ff5d80e6-ad76-49c5-a350-c7cd9e8528ae","Type":"ContainerStarted","Data":"d211eb6ad0193787133b5862a8603bbaec22408315fef0d1aff2f8646fe8ab1f"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.752754 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"26a1160122968b15f59c0d51ee5bc2dfbda73b4553b50358188d9fb09121eb43"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.752813 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"60c95c235fe6b0fbbf30cbfa2508be7d00a46542e2873f4514e2c9c4e2a94e80"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.763558 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" event={"ID":"3ae04268-6e5b-402f-b3c3-6b456df7e689","Type":"ContainerStarted","Data":"3812d624dcf788489f194699cf937047d8d590ddc598e6d3b6dfe5f2a921d535"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.765297 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cmrfr" event={"ID":"b423076b-47af-4e4e-a085-d47074f6bb8f","Type":"ContainerStarted","Data":"e6433b092535c812640dc689cf56485d6b1630245d6169bb85fa8ac75f2340c5"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.773676 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cj74l" event={"ID":"cbebf2b4-7952-45d6-8c77-efc6438e05fd","Type":"ContainerStarted","Data":"ba3904718b23c94a91b45864c35edb017d0ce452cc47efc039d671c3d144995a"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.783264 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" event={"ID":"9ada019a-2740-482e-864b-55762c330b26","Type":"ContainerStarted","Data":"de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.783339 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.783780 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:14 crc kubenswrapper[4954]: E1209 16:59:14.784963 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:15.284941664 +0000 UTC m=+151.673115484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.792684 4954 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6449l container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" start-of-body= Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.792740 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" podUID="9ada019a-2740-482e-864b-55762c330b26" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.795502 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" event={"ID":"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81","Type":"ContainerStarted","Data":"3ae33c49509f232f53e0192aa376275f6eeeac6124ea372e9313da830898fc77"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.816044 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" event={"ID":"192c8b71-0927-4047-8fa6-a0b45fdbef7b","Type":"ContainerStarted","Data":"42a5d4454dd441ca483014383b9241c6b4c1e1c017ef8aa8c452772efb454c71"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.817413 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.821037 4954 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-sspwb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.821097 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" podUID="192c8b71-0927-4047-8fa6-a0b45fdbef7b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.843665 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" event={"ID":"8f80de45-61f5-4538-93d7-c6252e3c64bc","Type":"ContainerStarted","Data":"0dbb26c9acceaf81c2da8c0a30e36479629544f1b82b2c1f7f128d7eab5bc285"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.878488 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" event={"ID":"454e49f6-b054-4901-bd18-461d7e9cd89e","Type":"ContainerStarted","Data":"4e238e14ac372d025d9b80ec7ad745692b7beb1abacb7e2aefa46d419d3e5184"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.885919 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:14 crc kubenswrapper[4954]: E1209 16:59:14.900435 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:15.400414182 +0000 UTC m=+151.788588202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.942251 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" event={"ID":"e43b0681-33e3-4eca-a737-ed3c9c2fe994","Type":"ContainerDied","Data":"0624ca1b677bf4f20c768f104b016c35e6d8fe8edd0d32ba07bc40c4cff6aa01"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.939732 4954 generic.go:334] "Generic (PLEG): container finished" podID="e43b0681-33e3-4eca-a737-ed3c9c2fe994" containerID="0624ca1b677bf4f20c768f104b016c35e6d8fe8edd0d32ba07bc40c4cff6aa01" exitCode=0 Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.973339 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" event={"ID":"05f2d085-d179-42d9-bc77-7dce694be634","Type":"ContainerStarted","Data":"eb46ebab93eaddbe8aea90b4b54225ca249939e6179c7f75e4bb9a7aa33a4aaa"} Dec 09 16:59:14 crc kubenswrapper[4954]: I1209 16:59:14.987896 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:14 crc kubenswrapper[4954]: E1209 16:59:14.989206 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:15.48918627 +0000 UTC m=+151.877360090 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.031178 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" event={"ID":"57dc805d-5ae9-4f59-80fb-0dd9e15a33e0","Type":"ContainerStarted","Data":"9ef6e3b580e8e798464eb45850687cefae4fbd7f711272f75aa96b3c8f3a9743"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.055421 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" podStartSLOduration=132.055378275 podStartE2EDuration="2m12.055378275s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:14.996791566 +0000 UTC m=+151.384965396" watchObservedRunningTime="2025-12-09 16:59:15.055378275 +0000 UTC m=+151.443552095" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.098562 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:15 crc kubenswrapper[4954]: E1209 16:59:15.101297 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:15.601269848 +0000 UTC m=+151.989443848 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.107367 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" event={"ID":"4fc4931a-8c26-4f29-8798-d730e10e8d39","Type":"ContainerStarted","Data":"c52d4ef10270827787a91d5ec9b495868e047056f36089aa0e7a886ea7d6cce2"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.125668 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" podStartSLOduration=132.124582307 podStartE2EDuration="2m12.124582307s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.122458635 +0000 UTC m=+151.510632475" watchObservedRunningTime="2025-12-09 16:59:15.124582307 +0000 UTC m=+151.512756127" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.127724 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" podStartSLOduration=132.127707006 podStartE2EDuration="2m12.127707006s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.087491106 +0000 UTC m=+151.475664926" watchObservedRunningTime="2025-12-09 16:59:15.127707006 +0000 UTC m=+151.515880836" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.148141 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" event={"ID":"df27667c-ecf7-4e33-9d34-c3a5b4088154","Type":"ContainerStarted","Data":"9f44a0bd785a5623866a26cfc3998f51050df68508d047f6ff752d7d0d99c13e"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.199790 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:15 crc kubenswrapper[4954]: E1209 16:59:15.200401 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:15.700365789 +0000 UTC m=+152.088539649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.200918 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:15 crc kubenswrapper[4954]: E1209 16:59:15.202473 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:15.702459541 +0000 UTC m=+152.090633361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.205829 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" event={"ID":"afae1fb2-041c-433d-9e7b-e29a7e82ed31","Type":"ContainerStarted","Data":"5282eb161b9911299e020613c5ea577923cfeb1af16d00a90cbe756ea32f9498"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.205894 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" event={"ID":"afae1fb2-041c-433d-9e7b-e29a7e82ed31","Type":"ContainerStarted","Data":"6468441a09faf8b0adf505cd3719b0c38dfb52436f34937803b65530bb71c722"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.215539 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" event={"ID":"f82b145c-6847-4a03-afff-2a513a79ecbe","Type":"ContainerStarted","Data":"b278bbcd42cb993a2068b8134b9ea7b140c32395d3309a9cd7e158cadd17f33f"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.215623 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" event={"ID":"f82b145c-6847-4a03-afff-2a513a79ecbe","Type":"ContainerStarted","Data":"da7f0854d1215b1d453ce017580eedf15a16548572e0b63836016becab2e7193"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.282827 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" event={"ID":"dd6facda-112d-4218-9cec-3e24d3d2249d","Type":"ContainerStarted","Data":"bdde1fe6aeb6736b12752b5713006dc2968ea0e5e587288bd558a2707621da87"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.283284 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" event={"ID":"dd6facda-112d-4218-9cec-3e24d3d2249d","Type":"ContainerStarted","Data":"30a2a9cd82091dd23fc511b49e08cc9c39712a7ecc665a65054fd884878c0c1f"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.306789 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" event={"ID":"4e939827-530f-4e2e-9455-080369e5c76d","Type":"ContainerStarted","Data":"fa2e1879ab2ec35d32857de302719a88b54e2bf580368578aef931825aeaff64"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.307041 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:15 crc kubenswrapper[4954]: E1209 16:59:15.307492 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:15.807470016 +0000 UTC m=+152.195643836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.310036 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" event={"ID":"ffebb621-9aa6-4104-a457-c180e6c93a28","Type":"ContainerStarted","Data":"d18c3d972fdd940e6adbbdd626f60125bbce701d748cc71ae410e85ded427864"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.339830 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.340382 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.346272 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" event={"ID":"1df4a2c0-8c86-447c-aa38-cb9a436f84eb","Type":"ContainerStarted","Data":"b53145e2c1186b379b30e2b5d575059fdf27a6e2912166dfeb591f874df7186f"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.361644 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" podStartSLOduration=132.361626239 podStartE2EDuration="2m12.361626239s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.359772338 +0000 UTC m=+151.747946168" watchObservedRunningTime="2025-12-09 16:59:15.361626239 +0000 UTC m=+151.749800059" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.363875 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" event={"ID":"9de03415-ecf6-43a2-8968-eae338108559","Type":"ContainerStarted","Data":"b85802697457e9d6bd7a76ade29b8baf4039defea4580fb249ca0a8525a4b635"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.365935 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.371492 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:15 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:15 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:15 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.371558 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.374457 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gw877" event={"ID":"eee9658e-289f-4ecf-8661-5b1bdf9c307c","Type":"ContainerStarted","Data":"c7a4e5c2d0ea905719e807ab88020733d56078757580717df185c1fc0417451d"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.418545 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" event={"ID":"a759536e-62a7-4410-b1a3-0ea059cd63f4","Type":"ContainerStarted","Data":"c7cd572c98f5f39d8d736b569f50573138fbec4815eb9da7a046c5b38fb82dad"} Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.428926 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:15 crc kubenswrapper[4954]: E1209 16:59:15.431972 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:15.931957808 +0000 UTC m=+152.320131618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.448822 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-cj74l" podStartSLOduration=133.448800828 podStartE2EDuration="2m13.448800828s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.391038504 +0000 UTC m=+151.779212324" watchObservedRunningTime="2025-12-09 16:59:15.448800828 +0000 UTC m=+151.836974648" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.470493 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" podStartSLOduration=133.470477827 podStartE2EDuration="2m13.470477827s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.46931612 +0000 UTC m=+151.857489940" watchObservedRunningTime="2025-12-09 16:59:15.470477827 +0000 UTC m=+151.858651647" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.530139 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:15 crc kubenswrapper[4954]: E1209 16:59:15.530395 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:16.030380564 +0000 UTC m=+152.418554384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.583566 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" podStartSLOduration=133.58355042 podStartE2EDuration="2m13.58355042s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.525715927 +0000 UTC m=+151.913889747" watchObservedRunningTime="2025-12-09 16:59:15.58355042 +0000 UTC m=+151.971724240" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.584050 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-48vd4" podStartSLOduration=133.584045643 podStartE2EDuration="2m13.584045643s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.581856571 +0000 UTC m=+151.970030391" watchObservedRunningTime="2025-12-09 16:59:15.584045643 +0000 UTC m=+151.972219463" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.635247 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:15 crc kubenswrapper[4954]: E1209 16:59:15.635623 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:16.13560438 +0000 UTC m=+152.523778200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.710267 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" podStartSLOduration=133.710251836 podStartE2EDuration="2m13.710251836s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.709987804 +0000 UTC m=+152.098161624" watchObservedRunningTime="2025-12-09 16:59:15.710251836 +0000 UTC m=+152.098425656" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.756812 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:15 crc kubenswrapper[4954]: E1209 16:59:15.757366 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:16.257346336 +0000 UTC m=+152.645520156 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.759729 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-chrbl" podStartSLOduration=132.75971826 podStartE2EDuration="2m12.75971826s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.758274192 +0000 UTC m=+152.146448012" watchObservedRunningTime="2025-12-09 16:59:15.75971826 +0000 UTC m=+152.147892070" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.888864 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-j9crx" podStartSLOduration=133.888836169 podStartE2EDuration="2m13.888836169s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.807014771 +0000 UTC m=+152.195188591" watchObservedRunningTime="2025-12-09 16:59:15.888836169 +0000 UTC m=+152.277009999" Dec 09 16:59:15 crc kubenswrapper[4954]: I1209 16:59:15.904453 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:15 crc kubenswrapper[4954]: E1209 16:59:15.904955 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:16.404940064 +0000 UTC m=+152.793113884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.009251 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:16 crc kubenswrapper[4954]: E1209 16:59:16.009555 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:16.509536618 +0000 UTC m=+152.897710438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.118479 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:16 crc kubenswrapper[4954]: E1209 16:59:16.119081 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:16.61906249 +0000 UTC m=+153.007236310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.214197 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" podStartSLOduration=133.214164787 podStartE2EDuration="2m13.214164787s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:15.984383278 +0000 UTC m=+152.372557098" watchObservedRunningTime="2025-12-09 16:59:16.214164787 +0000 UTC m=+152.602338607" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.214850 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nslc6" podStartSLOduration=133.21484464 podStartE2EDuration="2m13.21484464s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:16.203480923 +0000 UTC m=+152.591654753" watchObservedRunningTime="2025-12-09 16:59:16.21484464 +0000 UTC m=+152.603018460" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.222141 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:16 crc kubenswrapper[4954]: E1209 16:59:16.222429 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:16.722398236 +0000 UTC m=+153.110572056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.222700 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:16 crc kubenswrapper[4954]: E1209 16:59:16.223144 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:16.723129699 +0000 UTC m=+153.111303519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.339472 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:16 crc kubenswrapper[4954]: E1209 16:59:16.339850 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:16.839786245 +0000 UTC m=+153.227960195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.377010 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:16 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:16 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:16 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.377299 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.379857 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-96v5v" podStartSLOduration=134.379832094 podStartE2EDuration="2m14.379832094s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:16.265695303 +0000 UTC m=+152.653869123" watchObservedRunningTime="2025-12-09 16:59:16.379832094 +0000 UTC m=+152.768005914" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.421562 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-prclt" podStartSLOduration=134.421532131 podStartE2EDuration="2m14.421532131s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:16.419865741 +0000 UTC m=+152.808039551" watchObservedRunningTime="2025-12-09 16:59:16.421532131 +0000 UTC m=+152.809705951" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.431419 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" event={"ID":"4e939827-530f-4e2e-9455-080369e5c76d","Type":"ContainerStarted","Data":"8475f49ce4e7cc15f6aec00caa6d6b41f2b87de258783ec3888857b0a9443a51"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.431851 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.472849 4954 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qpkjl container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.472863 4954 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r7zd5 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.472910 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" podUID="db03f79f-fcea-4a92-b021-bac0b6c1cebc" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.472863 4954 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qpkjl container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.472975 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" podUID="db03f79f-fcea-4a92-b021-bac0b6c1cebc" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.472938 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" podUID="4e939827-530f-4e2e-9455-080369e5c76d" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.474049 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:16 crc kubenswrapper[4954]: E1209 16:59:16.474356 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:16.974345546 +0000 UTC m=+153.362519366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.534539 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" event={"ID":"05f2d085-d179-42d9-bc77-7dce694be634","Type":"ContainerStarted","Data":"b263428a89bf4cdaaf88774db18e3d86c1003a03769019431ae4bbe722b0fbf1"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.575378 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" event={"ID":"454e49f6-b054-4901-bd18-461d7e9cd89e","Type":"ContainerStarted","Data":"5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.576611 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.576889 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:16 crc kubenswrapper[4954]: E1209 16:59:16.578069 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:17.078050094 +0000 UTC m=+153.466223914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.580548 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" event={"ID":"4fc4931a-8c26-4f29-8798-d730e10e8d39","Type":"ContainerStarted","Data":"5934effc64bb2ea28a0f97a76a901c144dcffb5800b59f757b7ba572cb85b281"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.582371 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" event={"ID":"e43b0681-33e3-4eca-a737-ed3c9c2fe994","Type":"ContainerStarted","Data":"414dff1d4bbc957792237ed93390c079918d8f2727b7a49231970698c530b024"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.584107 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" event={"ID":"afae1fb2-041c-433d-9e7b-e29a7e82ed31","Type":"ContainerStarted","Data":"bdd852df9bbc1e014307a7b4ed7ae0782bcfd890a112b1815426b16613318648"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.584746 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.586139 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" event={"ID":"fcc2c29a-2e97-457a-8f25-d96a4aee4ea5","Type":"ContainerStarted","Data":"34d083351e05b5e44040a0a30f7769b375158551bec1fbf780c33f9a24029a16"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.587655 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" event={"ID":"192c8b71-0927-4047-8fa6-a0b45fdbef7b","Type":"ContainerStarted","Data":"46e8d4e49244fc98d3a6bdd8bb4a5560893397f6bb11efaab713d1137172178e"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.588502 4954 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-sspwb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.588647 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" podUID="192c8b71-0927-4047-8fa6-a0b45fdbef7b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.589941 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-qmwzp" event={"ID":"ffebb621-9aa6-4104-a457-c180e6c93a28","Type":"ContainerStarted","Data":"92dd3e2681b50a6cd0235c918f5183c2f7ac6a06a5a293bbb2de808046eff729"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.593272 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" event={"ID":"849c8fd6-0873-40d0-8c40-7bbc2365452d","Type":"ContainerStarted","Data":"068bca25c6f42bb4613f300cea36102d404f7f84460cd3aaf4bd1342ab4e4a80"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.594877 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" event={"ID":"7ec10355-df85-4f79-9bd5-78ff480e25f4","Type":"ContainerStarted","Data":"90cc45594ac7850f379585d2df9a5686ea111e6b20d2cf85bc4e63eec67e0ff5"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.596907 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" event={"ID":"3ae04268-6e5b-402f-b3c3-6b456df7e689","Type":"ContainerStarted","Data":"8897b9f99ad5d384ba936f5cb37e886794f009f63cbda8b108368e9f51fce7c3"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.597664 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.599383 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cmrfr" event={"ID":"b423076b-47af-4e4e-a085-d47074f6bb8f","Type":"ContainerStarted","Data":"c0cf2a8c5a216f7b265d4bf9b7393f8571b55754572cef502bf3872c8eb42bf9"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.601432 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gw877" event={"ID":"eee9658e-289f-4ecf-8661-5b1bdf9c307c","Type":"ContainerStarted","Data":"06af9dc7c8a1ee89b62ab3d2302411fb041b172ba75790151536f94489b56323"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.627956 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-q4bd2" podStartSLOduration=134.627938221 podStartE2EDuration="2m14.627938221s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:16.625897799 +0000 UTC m=+153.014071629" watchObservedRunningTime="2025-12-09 16:59:16.627938221 +0000 UTC m=+153.016112041" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.650724 4954 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8jkvw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.650779 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" podUID="454e49f6-b054-4901-bd18-461d7e9cd89e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.650853 4954 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pp8mx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.650868 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" podUID="3ae04268-6e5b-402f-b3c3-6b456df7e689" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.680998 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.681402 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8gzxc" podStartSLOduration=133.681391529 podStartE2EDuration="2m13.681391529s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:16.658821145 +0000 UTC m=+153.046994965" watchObservedRunningTime="2025-12-09 16:59:16.681391529 +0000 UTC m=+153.069565349" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.686463 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" podStartSLOduration=133.686450609 podStartE2EDuration="2m13.686450609s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:16.681103488 +0000 UTC m=+153.069277308" watchObservedRunningTime="2025-12-09 16:59:16.686450609 +0000 UTC m=+153.074624429" Dec 09 16:59:16 crc kubenswrapper[4954]: E1209 16:59:16.683309 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:17.18329637 +0000 UTC m=+153.571470190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.789168 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:16 crc kubenswrapper[4954]: E1209 16:59:16.789663 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:17.289638114 +0000 UTC m=+153.677811934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.837803 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7jpc4" event={"ID":"8f80de45-61f5-4538-93d7-c6252e3c64bc","Type":"ContainerStarted","Data":"8a0b1ceb1cd78cd2823ddb1a61cc7646b5139a206c9e56950dcb4d201e089ad0"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.868903 4954 generic.go:334] "Generic (PLEG): container finished" podID="c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81" containerID="3ae33c49509f232f53e0192aa376275f6eeeac6124ea372e9313da830898fc77" exitCode=0 Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.869139 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" event={"ID":"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81","Type":"ContainerDied","Data":"3ae33c49509f232f53e0192aa376275f6eeeac6124ea372e9313da830898fc77"} Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.874845 4954 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-qpkjl container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.875137 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" podUID="db03f79f-fcea-4a92-b021-bac0b6c1cebc" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.894355 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:16 crc kubenswrapper[4954]: E1209 16:59:16.894796 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:17.39477385 +0000 UTC m=+153.782947670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.903886 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-cmrfr" podStartSLOduration=8.903864854 podStartE2EDuration="8.903864854s" podCreationTimestamp="2025-12-09 16:59:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:16.701972711 +0000 UTC m=+153.090146551" watchObservedRunningTime="2025-12-09 16:59:16.903864854 +0000 UTC m=+153.292038674" Dec 09 16:59:16 crc kubenswrapper[4954]: I1209 16:59:16.958153 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.004031 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.005160 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:17.505140187 +0000 UTC m=+153.893314007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.051861 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mhkmt" podStartSLOduration=135.051837035 podStartE2EDuration="2m15.051837035s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:16.995185978 +0000 UTC m=+153.383359798" watchObservedRunningTime="2025-12-09 16:59:17.051837035 +0000 UTC m=+153.440010855" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.052719 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zlqrk" podStartSLOduration=135.05271446 podStartE2EDuration="2m15.05271446s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:17.052153167 +0000 UTC m=+153.440326987" watchObservedRunningTime="2025-12-09 16:59:17.05271446 +0000 UTC m=+153.440888270" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.137455 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.137938 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:17.637921318 +0000 UTC m=+154.026095138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.153139 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.258207 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.259071 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:17.759056159 +0000 UTC m=+154.147229979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.260732 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" podStartSLOduration=134.26071842 podStartE2EDuration="2m14.26071842s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:17.153218649 +0000 UTC m=+153.541392489" watchObservedRunningTime="2025-12-09 16:59:17.26071842 +0000 UTC m=+153.648892240" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.261195 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" podStartSLOduration=134.261187622 podStartE2EDuration="2m14.261187622s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:17.258159774 +0000 UTC m=+153.646333594" watchObservedRunningTime="2025-12-09 16:59:17.261187622 +0000 UTC m=+153.649361442" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.361647 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.361962 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:17.861951133 +0000 UTC m=+154.250124953 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.417392 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.442449 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:17 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:17 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:17 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.442571 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.473052 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.473308 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:17.973292096 +0000 UTC m=+154.361465916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.563735 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" podStartSLOduration=134.563715124 podStartE2EDuration="2m14.563715124s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:17.360098831 +0000 UTC m=+153.748272661" watchObservedRunningTime="2025-12-09 16:59:17.563715124 +0000 UTC m=+153.951888944" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.576222 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.576564 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.07655203 +0000 UTC m=+154.464725850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.677155 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.677380 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.177345421 +0000 UTC m=+154.565519251 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.677509 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.677844 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.177830284 +0000 UTC m=+154.566004104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.785051 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.785307 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.285292853 +0000 UTC m=+154.673466673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.886065 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.886444 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.386433486 +0000 UTC m=+154.774607306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.929766 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" event={"ID":"fcc2c29a-2e97-457a-8f25-d96a4aee4ea5","Type":"ContainerStarted","Data":"48c18e3541e3b597379610000ba07ca9fc0534c9c4b79bea6e81d2aaf50a9941"} Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.931914 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" event={"ID":"849c8fd6-0873-40d0-8c40-7bbc2365452d","Type":"ContainerStarted","Data":"c4e5db9841ffe98990c1a117579a35f873a19c76f88fb2e60147d2a1524500e4"} Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.956323 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" event={"ID":"e43b0681-33e3-4eca-a737-ed3c9c2fe994","Type":"ContainerStarted","Data":"7ffdaa777dbc5fc55814c377c23a2723afdda7cc20e3e67acc57f890485df6ce"} Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.959331 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gw877" event={"ID":"eee9658e-289f-4ecf-8661-5b1bdf9c307c","Type":"ContainerStarted","Data":"c5698eb2ff75d8708b5a20ad191b7b601c9154625d5c02e1d94a5444235b9312"} Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.959383 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-gw877" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.961923 4954 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-sspwb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.962055 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" podUID="192c8b71-0927-4047-8fa6-a0b45fdbef7b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.962287 4954 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r7zd5 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.962380 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" podUID="4e939827-530f-4e2e-9455-080369e5c76d" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.962402 4954 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8jkvw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.962455 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" podUID="454e49f6-b054-4901-bd18-461d7e9cd89e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.963092 4954 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pp8mx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.963129 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" podUID="3ae04268-6e5b-402f-b3c3-6b456df7e689" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.987844 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:17 crc kubenswrapper[4954]: E1209 16:59:17.990336 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.490315354 +0000 UTC m=+154.878489174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:17 crc kubenswrapper[4954]: I1209 16:59:17.991830 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.057384 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.557366714 +0000 UTC m=+154.945540534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.099132 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.110325 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.610294409 +0000 UTC m=+154.998468229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.179867 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" podStartSLOduration=136.179850664 podStartE2EDuration="2m16.179850664s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:18.177248358 +0000 UTC m=+154.565422178" watchObservedRunningTime="2025-12-09 16:59:18.179850664 +0000 UTC m=+154.568024484" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.180339 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-pz4hd" podStartSLOduration=135.180333466 podStartE2EDuration="2m15.180333466s" podCreationTimestamp="2025-12-09 16:57:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:18.074012193 +0000 UTC m=+154.462186013" watchObservedRunningTime="2025-12-09 16:59:18.180333466 +0000 UTC m=+154.568507286" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.201845 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.202172 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.702160777 +0000 UTC m=+155.090334587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.206865 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hp66s" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.206916 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.207630 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.240843 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.241434 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.243098 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.306479 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-gw877" podStartSLOduration=10.306457297 podStartE2EDuration="10.306457297s" podCreationTimestamp="2025-12-09 16:59:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:18.301495808 +0000 UTC m=+154.689669628" watchObservedRunningTime="2025-12-09 16:59:18.306457297 +0000 UTC m=+154.694631117" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.320839 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.322963 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.822900315 +0000 UTC m=+155.211074135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.332135 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1edf562-5b72-4ac1-b801-29226c3c43c8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e1edf562-5b72-4ac1-b801-29226c3c43c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.332311 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1edf562-5b72-4ac1-b801-29226c3c43c8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e1edf562-5b72-4ac1-b801-29226c3c43c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.332470 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.333150 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.833126146 +0000 UTC m=+155.221299966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.426654 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:18 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:18 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:18 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.426718 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.433870 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.434109 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1edf562-5b72-4ac1-b801-29226c3c43c8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e1edf562-5b72-4ac1-b801-29226c3c43c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.434158 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1edf562-5b72-4ac1-b801-29226c3c43c8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e1edf562-5b72-4ac1-b801-29226c3c43c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.434257 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1edf562-5b72-4ac1-b801-29226c3c43c8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e1edf562-5b72-4ac1-b801-29226c3c43c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.434339 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:18.93432278 +0000 UTC m=+155.322496600 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.434834 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lz9lt" podStartSLOduration=136.434814962 podStartE2EDuration="2m16.434814962s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:18.433210393 +0000 UTC m=+154.821384213" watchObservedRunningTime="2025-12-09 16:59:18.434814962 +0000 UTC m=+154.822988782" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.573656 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.574046 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:19.074033792 +0000 UTC m=+155.462207612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.587542 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1edf562-5b72-4ac1-b801-29226c3c43c8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e1edf562-5b72-4ac1-b801-29226c3c43c8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.622446 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.675448 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.675802 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:19.175772397 +0000 UTC m=+155.563946217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.675991 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.676346 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:19.17633307 +0000 UTC m=+155.564506900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.783893 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.784291 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:19.284272014 +0000 UTC m=+155.672445834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.848967 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-qpkjl" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.939696 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:18 crc kubenswrapper[4954]: E1209 16:59:18.940138 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:19.440120322 +0000 UTC m=+155.828294142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.979292 4954 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8jkvw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.979341 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" podUID="454e49f6-b054-4901-bd18-461d7e9cd89e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.979662 4954 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pp8mx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Dec 09 16:59:18 crc kubenswrapper[4954]: I1209 16:59:18.979693 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" podUID="3ae04268-6e5b-402f-b3c3-6b456df7e689" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.040207 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:19 crc kubenswrapper[4954]: E1209 16:59:19.041773 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:19.541753227 +0000 UTC m=+155.929927047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.152955 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:19 crc kubenswrapper[4954]: E1209 16:59:19.153273 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:19.653262431 +0000 UTC m=+156.041436251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.264568 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:19 crc kubenswrapper[4954]: E1209 16:59:19.264872 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:19.764852456 +0000 UTC m=+156.153026276 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.391606 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:19 crc kubenswrapper[4954]: E1209 16:59:19.391899 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:19.891887053 +0000 UTC m=+156.280060863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.500561 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:19 crc kubenswrapper[4954]: E1209 16:59:19.500893 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.000861151 +0000 UTC m=+156.389034981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.501136 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:19 crc kubenswrapper[4954]: E1209 16:59:19.501504 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.001487194 +0000 UTC m=+156.389661014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.601724 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:19 crc kubenswrapper[4954]: E1209 16:59:19.601899 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.101871863 +0000 UTC m=+156.490045683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.602104 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:19 crc kubenswrapper[4954]: E1209 16:59:19.602420 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.102412446 +0000 UTC m=+156.490586266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.606664 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:19 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:19 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:19 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.606714 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.651707 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.652270 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.654886 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.655183 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.802606 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.802754 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"981f3a4c-bcf3-4a51-af62-28a3f65b7370\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.802773 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"981f3a4c-bcf3-4a51-af62-28a3f65b7370\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 16:59:19 crc kubenswrapper[4954]: E1209 16:59:19.802917 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.3029 +0000 UTC m=+156.691073820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.941195 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.941274 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"981f3a4c-bcf3-4a51-af62-28a3f65b7370\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.941292 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"981f3a4c-bcf3-4a51-af62-28a3f65b7370\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.941622 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"981f3a4c-bcf3-4a51-af62-28a3f65b7370\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 16:59:19 crc kubenswrapper[4954]: E1209 16:59:19.941679 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.441662856 +0000 UTC m=+156.829836676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:19 crc kubenswrapper[4954]: I1209 16:59:19.943616 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.042534 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:20 crc kubenswrapper[4954]: E1209 16:59:20.042749 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.542720088 +0000 UTC m=+156.930893908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.043192 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:20 crc kubenswrapper[4954]: E1209 16:59:20.044856 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.544844561 +0000 UTC m=+156.933018381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.059542 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"981f3a4c-bcf3-4a51-af62-28a3f65b7370\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.166108 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:20 crc kubenswrapper[4954]: E1209 16:59:20.166509 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.666491566 +0000 UTC m=+157.054665386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.267334 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:20 crc kubenswrapper[4954]: E1209 16:59:20.267729 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.767717409 +0000 UTC m=+157.155891229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.344814 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.368961 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:20 crc kubenswrapper[4954]: E1209 16:59:20.369378 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.869362724 +0000 UTC m=+157.257536544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.405338 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.405392 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.405832 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:20 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:20 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:20 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.405856 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.405940 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.405960 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.480770 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:20 crc kubenswrapper[4954]: E1209 16:59:20.481153 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:20.981141529 +0000 UTC m=+157.369315349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.666752 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.667731 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:20 crc kubenswrapper[4954]: E1209 16:59:20.668126 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:21.168096453 +0000 UTC m=+157.556270263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.668934 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.674636 4954 patch_prober.go:28] interesting pod/console-f9d7485db-v4pm7 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.39:8443/health\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.674694 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-v4pm7" podUID="7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.39:8443/health\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.780517 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:20 crc kubenswrapper[4954]: E1209 16:59:20.782311 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:21.282298183 +0000 UTC m=+157.670472003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.848779 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.884102 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:20 crc kubenswrapper[4954]: E1209 16:59:20.884470 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:21.384454041 +0000 UTC m=+157.772627861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.888521 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.987151 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-config-volume\") pod \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.987223 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-secret-volume\") pod \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.990356 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-config-volume" (OuterVolumeSpecName: "config-volume") pod "c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81" (UID: "c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:59:20 crc kubenswrapper[4954]: I1209 16:59:20.994779 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bfpg\" (UniqueName: \"kubernetes.io/projected/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-kube-api-access-6bfpg\") pod \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\" (UID: \"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81\") " Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.127277 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.128127 4954 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.128630 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:21.628609015 +0000 UTC m=+158.016782835 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.134670 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.134706 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.179584 4954 patch_prober.go:28] interesting pod/apiserver-76f77b778f-v6ftv container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.179650 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" podUID="e43b0681-33e3-4eca-a737-ed3c9c2fe994" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.191345 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81" (UID: "c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.242226 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.242579 4954 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.243293 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:21.743275258 +0000 UTC m=+158.131449078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.254607 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-kube-api-access-6bfpg" (OuterVolumeSpecName: "kube-api-access-6bfpg") pod "c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81" (UID: "c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81"). InnerVolumeSpecName "kube-api-access-6bfpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.338943 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" event={"ID":"c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81","Type":"ContainerDied","Data":"d555a1e11e57675734c870b16bb4afbee52903dc26ab5e8a1feaa1df43e0d91c"} Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.338994 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d555a1e11e57675734c870b16bb4afbee52903dc26ab5e8a1feaa1df43e0d91c" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.339108 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.350011 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.350108 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bfpg\" (UniqueName: \"kubernetes.io/projected/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81-kube-api-access-6bfpg\") on node \"crc\" DevicePath \"\"" Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.350425 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:21.850410137 +0000 UTC m=+158.238583957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.351042 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e1edf562-5b72-4ac1-b801-29226c3c43c8","Type":"ContainerStarted","Data":"99a263b04142c01ce121ba777ff3eea2d49e7758d6c926c4699a92a30b0fbf69"} Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.352997 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" event={"ID":"4fc4931a-8c26-4f29-8798-d730e10e8d39","Type":"ContainerStarted","Data":"187c95d27cc3dbe7d05e966fa3a4354bc69f42be13fbe4eac843afa0a1b3d9fc"} Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.367841 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.382144 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:21 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:21 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:21 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.382197 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.421587 4954 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8jkvw container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.421647 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" podUID="454e49f6-b054-4901-bd18-461d7e9cd89e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.421749 4954 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8jkvw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.421852 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" podUID="454e49f6-b054-4901-bd18-461d7e9cd89e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.436448 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r7zd5" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.450812 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.451065 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:21.951039526 +0000 UTC m=+158.339213346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.451143 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.451496 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:21.951484609 +0000 UTC m=+158.339658429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.463051 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sspwb" Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.551730 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.551907 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:22.051883356 +0000 UTC m=+158.440057176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.553827 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.554986 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:22.054976845 +0000 UTC m=+158.443150665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.685783 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.688250 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:22.188230818 +0000 UTC m=+158.576404648 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.790444 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.790812 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:22.29079832 +0000 UTC m=+158.678972140 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.891502 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.894401 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:22.394376207 +0000 UTC m=+158.782550027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:21 crc kubenswrapper[4954]: I1209 16:59:21.996512 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:21 crc kubenswrapper[4954]: E1209 16:59:21.996850 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:22.496837397 +0000 UTC m=+158.885011217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.081142 4954 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.173498 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:22 crc kubenswrapper[4954]: E1209 16:59:22.174190 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:22.674149433 +0000 UTC m=+159.062323253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.274783 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:22 crc kubenswrapper[4954]: E1209 16:59:22.275231 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 16:59:22.775214835 +0000 UTC m=+159.163388655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dx5bf" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.358651 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e1edf562-5b72-4ac1-b801-29226c3c43c8","Type":"ContainerStarted","Data":"1fc4047e0d4e7ab455550162236f62285ff06046d134bfbc1614787f3517a904"} Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.360921 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" event={"ID":"4fc4931a-8c26-4f29-8798-d730e10e8d39","Type":"ContainerStarted","Data":"d3af0dda8ef5394cdb1e869b4b7d01771af572d83d55a5f7609c8e653ee49335"} Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.372486 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:22 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:22 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:22 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.372556 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.375733 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:22 crc kubenswrapper[4954]: E1209 16:59:22.376299 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 16:59:22.876281107 +0000 UTC m=+159.264454927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.421422 4954 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-09T16:59:22.081167439Z","Handler":null,"Name":""} Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.426854 4954 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.426890 4954 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.431559 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.431543686 podStartE2EDuration="4.431543686s" podCreationTimestamp="2025-12-09 16:59:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:22.430941972 +0000 UTC m=+158.819115792" watchObservedRunningTime="2025-12-09 16:59:22.431543686 +0000 UTC m=+158.819717506" Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.476865 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.484688 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 16:59:22 crc kubenswrapper[4954]: W1209 16:59:22.495754 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod981f3a4c_bcf3_4a51_af62_28a3f65b7370.slice/crio-ac50edf9af0ce63241997d865caa3fdf2fa64b84b2b030d8697a533d36a0dccb WatchSource:0}: Error finding container ac50edf9af0ce63241997d865caa3fdf2fa64b84b2b030d8697a533d36a0dccb: Status 404 returned error can't find the container with id ac50edf9af0ce63241997d865caa3fdf2fa64b84b2b030d8697a533d36a0dccb Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.541568 4954 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pp8mx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.541965 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" podUID="3ae04268-6e5b-402f-b3c3-6b456df7e689" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.542344 4954 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pp8mx container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 16:59:22 crc kubenswrapper[4954]: I1209 16:59:22.542365 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" podUID="3ae04268-6e5b-402f-b3c3-6b456df7e689" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.112305 4954 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.112367 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.227246 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dx5bf\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.337009 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.338026 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.342629 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.376674 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:23 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:23 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:23 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.376770 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.386441 4954 generic.go:334] "Generic (PLEG): container finished" podID="e1edf562-5b72-4ac1-b801-29226c3c43c8" containerID="1fc4047e0d4e7ab455550162236f62285ff06046d134bfbc1614787f3517a904" exitCode=0 Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.386563 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e1edf562-5b72-4ac1-b801-29226c3c43c8","Type":"ContainerDied","Data":"1fc4047e0d4e7ab455550162236f62285ff06046d134bfbc1614787f3517a904"} Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.405446 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" event={"ID":"4fc4931a-8c26-4f29-8798-d730e10e8d39","Type":"ContainerStarted","Data":"3634e668b2b0eea7c045edf54c868c08cf59a957a4e0f86aebc5ab58c1a69298"} Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.430943 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"981f3a4c-bcf3-4a51-af62-28a3f65b7370","Type":"ContainerStarted","Data":"ab5a8118f63c4c7931595ccce783ae3eadb46bc15db93a51cedd400b130b1fd0"} Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.430985 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"981f3a4c-bcf3-4a51-af62-28a3f65b7370","Type":"ContainerStarted","Data":"ac50edf9af0ce63241997d865caa3fdf2fa64b84b2b030d8697a533d36a0dccb"} Dec 09 16:59:23 crc kubenswrapper[4954]: I1209 16:59:23.493392 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-b4gqt" podStartSLOduration=15.49337574 podStartE2EDuration="15.49337574s" podCreationTimestamp="2025-12-09 16:59:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:23.44801314 +0000 UTC m=+159.836186980" watchObservedRunningTime="2025-12-09 16:59:23.49337574 +0000 UTC m=+159.881549560" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.002619 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=5.002585843 podStartE2EDuration="5.002585843s" podCreationTimestamp="2025-12-09 16:59:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:23.513778501 +0000 UTC m=+159.901952341" watchObservedRunningTime="2025-12-09 16:59:24.002585843 +0000 UTC m=+160.390759663" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.004844 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dx5bf"] Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.150771 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.622850 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.626674 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:24 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:24 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:24 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.627144 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.630730 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/856fc314-24bd-403e-a4b0-3dcd73eba595-metrics-certs\") pod \"network-metrics-daemon-ffkzn\" (UID: \"856fc314-24bd-403e-a4b0-3dcd73eba595\") " pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.660122 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" event={"ID":"d284ca9d-999e-4ab5-8d7e-647af799d69c","Type":"ContainerStarted","Data":"04e8fd59b0bd3bb85eea438d9c0e8b48560d6ba53e15e3d073801b4529d045bb"} Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.727902 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-snfrc"] Dec 09 16:59:24 crc kubenswrapper[4954]: E1209 16:59:24.728296 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81" containerName="collect-profiles" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.728323 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81" containerName="collect-profiles" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.728474 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81" containerName="collect-profiles" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.730271 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.730542 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nlf5c"] Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.731855 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.735510 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.735757 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.818451 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-snfrc"] Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.822203 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nlf5c"] Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.824123 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-catalog-content\") pod \"community-operators-snfrc\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.824165 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pzhr\" (UniqueName: \"kubernetes.io/projected/271559e2-f374-4aeb-a3f7-6b039afc7e8f-kube-api-access-9pzhr\") pod \"certified-operators-nlf5c\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.824201 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-utilities\") pod \"certified-operators-nlf5c\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.824216 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-catalog-content\") pod \"certified-operators-nlf5c\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.824234 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-utilities\") pod \"community-operators-snfrc\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.824253 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zdx7\" (UniqueName: \"kubernetes.io/projected/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-kube-api-access-2zdx7\") pod \"community-operators-snfrc\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.857006 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ffkzn" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.877652 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j57d6"] Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.879210 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.898162 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j57d6"] Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.924956 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pzhr\" (UniqueName: \"kubernetes.io/projected/271559e2-f374-4aeb-a3f7-6b039afc7e8f-kube-api-access-9pzhr\") pod \"certified-operators-nlf5c\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.924993 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmdfn\" (UniqueName: \"kubernetes.io/projected/f28a7275-4712-4dc9-89a3-16a77bba72d1-kube-api-access-nmdfn\") pod \"community-operators-j57d6\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925040 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-utilities\") pod \"certified-operators-nlf5c\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925055 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-catalog-content\") pod \"certified-operators-nlf5c\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925408 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-utilities\") pod \"certified-operators-nlf5c\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925074 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-utilities\") pod \"community-operators-snfrc\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925473 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zdx7\" (UniqueName: \"kubernetes.io/projected/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-kube-api-access-2zdx7\") pod \"community-operators-snfrc\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925484 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-catalog-content\") pod \"certified-operators-nlf5c\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925497 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-catalog-content\") pod \"community-operators-j57d6\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925544 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-utilities\") pod \"community-operators-j57d6\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925577 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-catalog-content\") pod \"community-operators-snfrc\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925681 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-utilities\") pod \"community-operators-snfrc\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.925905 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-catalog-content\") pod \"community-operators-snfrc\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.942649 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zdx7\" (UniqueName: \"kubernetes.io/projected/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-kube-api-access-2zdx7\") pod \"community-operators-snfrc\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.943825 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pzhr\" (UniqueName: \"kubernetes.io/projected/271559e2-f374-4aeb-a3f7-6b039afc7e8f-kube-api-access-9pzhr\") pod \"certified-operators-nlf5c\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.970489 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jjxbl"] Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.971697 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:24 crc kubenswrapper[4954]: I1209 16:59:24.984318 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jjxbl"] Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.047250 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlcll\" (UniqueName: \"kubernetes.io/projected/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-kube-api-access-qlcll\") pod \"certified-operators-jjxbl\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.047332 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmdfn\" (UniqueName: \"kubernetes.io/projected/f28a7275-4712-4dc9-89a3-16a77bba72d1-kube-api-access-nmdfn\") pod \"community-operators-j57d6\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.047376 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-catalog-content\") pod \"certified-operators-jjxbl\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.047424 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-catalog-content\") pod \"community-operators-j57d6\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.047448 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-utilities\") pod \"certified-operators-jjxbl\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.047466 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-utilities\") pod \"community-operators-j57d6\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.048348 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-utilities\") pod \"community-operators-j57d6\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.048997 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-catalog-content\") pod \"community-operators-j57d6\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.063743 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmdfn\" (UniqueName: \"kubernetes.io/projected/f28a7275-4712-4dc9-89a3-16a77bba72d1-kube-api-access-nmdfn\") pod \"community-operators-j57d6\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.188642 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-snfrc" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.188991 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.189372 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-utilities\") pod \"certified-operators-jjxbl\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.188946 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-utilities\") pod \"certified-operators-jjxbl\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.189472 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlcll\" (UniqueName: \"kubernetes.io/projected/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-kube-api-access-qlcll\") pod \"certified-operators-jjxbl\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.189501 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-catalog-content\") pod \"certified-operators-jjxbl\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.189799 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-catalog-content\") pod \"certified-operators-jjxbl\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.207951 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlcll\" (UniqueName: \"kubernetes.io/projected/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-kube-api-access-qlcll\") pod \"certified-operators-jjxbl\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.227954 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j57d6" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.289473 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.382969 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:25 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:25 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:25 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.383026 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:25 crc kubenswrapper[4954]: I1209 16:59:25.648067 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ffkzn"] Dec 09 16:59:26 crc kubenswrapper[4954]: W1209 16:59:26.088514 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod856fc314_24bd_403e_a4b0_3dcd73eba595.slice/crio-59a2760a7bf9f5c32085800ba1f718ed0cde92d8292cfad269d8de90c49a3e9e WatchSource:0}: Error finding container 59a2760a7bf9f5c32085800ba1f718ed0cde92d8292cfad269d8de90c49a3e9e: Status 404 returned error can't find the container with id 59a2760a7bf9f5c32085800ba1f718ed0cde92d8292cfad269d8de90c49a3e9e Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.189179 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.323587 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1edf562-5b72-4ac1-b801-29226c3c43c8-kubelet-dir\") pod \"e1edf562-5b72-4ac1-b801-29226c3c43c8\" (UID: \"e1edf562-5b72-4ac1-b801-29226c3c43c8\") " Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.324047 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1edf562-5b72-4ac1-b801-29226c3c43c8-kube-api-access\") pod \"e1edf562-5b72-4ac1-b801-29226c3c43c8\" (UID: \"e1edf562-5b72-4ac1-b801-29226c3c43c8\") " Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.323696 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1edf562-5b72-4ac1-b801-29226c3c43c8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e1edf562-5b72-4ac1-b801-29226c3c43c8" (UID: "e1edf562-5b72-4ac1-b801-29226c3c43c8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.324279 4954 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1edf562-5b72-4ac1-b801-29226c3c43c8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.329710 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1edf562-5b72-4ac1-b801-29226c3c43c8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e1edf562-5b72-4ac1-b801-29226c3c43c8" (UID: "e1edf562-5b72-4ac1-b801-29226c3c43c8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.383438 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:26 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:26 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:26 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.383491 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.425940 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1edf562-5b72-4ac1-b801-29226c3c43c8-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.563908 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7mwfq"] Dec 09 16:59:26 crc kubenswrapper[4954]: E1209 16:59:26.564123 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1edf562-5b72-4ac1-b801-29226c3c43c8" containerName="pruner" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.564136 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1edf562-5b72-4ac1-b801-29226c3c43c8" containerName="pruner" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.564229 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1edf562-5b72-4ac1-b801-29226c3c43c8" containerName="pruner" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.565040 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.571044 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.572694 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7mwfq"] Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.628012 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-catalog-content\") pod \"redhat-marketplace-7mwfq\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.628164 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-utilities\") pod \"redhat-marketplace-7mwfq\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.628195 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjj6v\" (UniqueName: \"kubernetes.io/projected/bcc59fe2-e410-4646-be11-560d0f45938b-kube-api-access-sjj6v\") pod \"redhat-marketplace-7mwfq\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.687947 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-gw877" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.697046 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" event={"ID":"856fc314-24bd-403e-a4b0-3dcd73eba595","Type":"ContainerStarted","Data":"59a2760a7bf9f5c32085800ba1f718ed0cde92d8292cfad269d8de90c49a3e9e"} Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.699195 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e1edf562-5b72-4ac1-b801-29226c3c43c8","Type":"ContainerDied","Data":"99a263b04142c01ce121ba777ff3eea2d49e7758d6c926c4699a92a30b0fbf69"} Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.699221 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99a263b04142c01ce121ba777ff3eea2d49e7758d6c926c4699a92a30b0fbf69" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.699271 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.733370 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-utilities\") pod \"redhat-marketplace-7mwfq\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.733755 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjj6v\" (UniqueName: \"kubernetes.io/projected/bcc59fe2-e410-4646-be11-560d0f45938b-kube-api-access-sjj6v\") pod \"redhat-marketplace-7mwfq\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.733850 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-catalog-content\") pod \"redhat-marketplace-7mwfq\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.736373 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-catalog-content\") pod \"redhat-marketplace-7mwfq\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.736331 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-utilities\") pod \"redhat-marketplace-7mwfq\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.869910 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjj6v\" (UniqueName: \"kubernetes.io/projected/bcc59fe2-e410-4646-be11-560d0f45938b-kube-api-access-sjj6v\") pod \"redhat-marketplace-7mwfq\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.880097 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.919691 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-snfrc"] Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.966984 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c9xz9"] Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.967983 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:26 crc kubenswrapper[4954]: I1209 16:59:26.977965 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9xz9"] Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.003162 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nlf5c"] Dec 09 16:59:27 crc kubenswrapper[4954]: W1209 16:59:27.007873 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod271559e2_f374_4aeb_a3f7_6b039afc7e8f.slice/crio-b28cf9b500bb266e35dbafe339f173310868e786e6f62ba93da51fc5486d8342 WatchSource:0}: Error finding container b28cf9b500bb266e35dbafe339f173310868e786e6f62ba93da51fc5486d8342: Status 404 returned error can't find the container with id b28cf9b500bb266e35dbafe339f173310868e786e6f62ba93da51fc5486d8342 Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.038757 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjprj\" (UniqueName: \"kubernetes.io/projected/f344130d-4bc2-4575-bb60-815ea6ae6a84-kube-api-access-sjprj\") pod \"redhat-marketplace-c9xz9\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.038819 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-utilities\") pod \"redhat-marketplace-c9xz9\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.038849 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-catalog-content\") pod \"redhat-marketplace-c9xz9\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.140610 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjprj\" (UniqueName: \"kubernetes.io/projected/f344130d-4bc2-4575-bb60-815ea6ae6a84-kube-api-access-sjprj\") pod \"redhat-marketplace-c9xz9\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.140685 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-utilities\") pod \"redhat-marketplace-c9xz9\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.140729 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-catalog-content\") pod \"redhat-marketplace-c9xz9\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.141391 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-catalog-content\") pod \"redhat-marketplace-c9xz9\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.141913 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-utilities\") pod \"redhat-marketplace-c9xz9\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.157292 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jjxbl"] Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.170633 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjprj\" (UniqueName: \"kubernetes.io/projected/f344130d-4bc2-4575-bb60-815ea6ae6a84-kube-api-access-sjprj\") pod \"redhat-marketplace-c9xz9\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.173388 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j57d6"] Dec 09 16:59:27 crc kubenswrapper[4954]: W1209 16:59:27.203947 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf28a7275_4712_4dc9_89a3_16a77bba72d1.slice/crio-8dcea5cde4292cf3e3b5f0ecf7804687718f4456131e9b2d7bbf191afba8827b WatchSource:0}: Error finding container 8dcea5cde4292cf3e3b5f0ecf7804687718f4456131e9b2d7bbf191afba8827b: Status 404 returned error can't find the container with id 8dcea5cde4292cf3e3b5f0ecf7804687718f4456131e9b2d7bbf191afba8827b Dec 09 16:59:27 crc kubenswrapper[4954]: W1209 16:59:27.204408 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9ebbd26_bcdc_4db3_b345_a5715ac59a82.slice/crio-ded987bb7da373d6123e13a47a53040e806b2aaf10006f0503aacf00428712ab WatchSource:0}: Error finding container ded987bb7da373d6123e13a47a53040e806b2aaf10006f0503aacf00428712ab: Status 404 returned error can't find the container with id ded987bb7da373d6123e13a47a53040e806b2aaf10006f0503aacf00428712ab Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.231084 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7mwfq"] Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.295650 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.369394 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:27 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:27 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:27 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.369449 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.509903 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9xz9"] Dec 09 16:59:27 crc kubenswrapper[4954]: W1209 16:59:27.516239 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf344130d_4bc2_4575_bb60_815ea6ae6a84.slice/crio-2cd1325e49b8a5e2d271d52cd787123c775adf449b5ac0b59c57a02b9bd05874 WatchSource:0}: Error finding container 2cd1325e49b8a5e2d271d52cd787123c775adf449b5ac0b59c57a02b9bd05874: Status 404 returned error can't find the container with id 2cd1325e49b8a5e2d271d52cd787123c775adf449b5ac0b59c57a02b9bd05874 Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.566007 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vfz9w"] Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.566977 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.570684 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.585988 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vfz9w"] Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.646814 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mnvn\" (UniqueName: \"kubernetes.io/projected/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-kube-api-access-7mnvn\") pod \"redhat-operators-vfz9w\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.646871 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-catalog-content\") pod \"redhat-operators-vfz9w\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.647004 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-utilities\") pod \"redhat-operators-vfz9w\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.704996 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nlf5c" event={"ID":"271559e2-f374-4aeb-a3f7-6b039afc7e8f","Type":"ContainerStarted","Data":"b28cf9b500bb266e35dbafe339f173310868e786e6f62ba93da51fc5486d8342"} Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.706407 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-snfrc" event={"ID":"1df4d5f7-03d7-4b91-b949-b555dae7ce2b","Type":"ContainerStarted","Data":"66b3195cf75c18ea7476f3b74fdc85ae98b35d7058605fe6b7e367dca7ebae85"} Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.707954 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9xz9" event={"ID":"f344130d-4bc2-4575-bb60-815ea6ae6a84","Type":"ContainerStarted","Data":"2cd1325e49b8a5e2d271d52cd787123c775adf449b5ac0b59c57a02b9bd05874"} Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.708791 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7mwfq" event={"ID":"bcc59fe2-e410-4646-be11-560d0f45938b","Type":"ContainerStarted","Data":"bcaebf94cf10f78527abfeb2876fc1f449ff6538cc4ca8f43e4a24a704296945"} Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.710066 4954 generic.go:334] "Generic (PLEG): container finished" podID="981f3a4c-bcf3-4a51-af62-28a3f65b7370" containerID="ab5a8118f63c4c7931595ccce783ae3eadb46bc15db93a51cedd400b130b1fd0" exitCode=0 Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.710114 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"981f3a4c-bcf3-4a51-af62-28a3f65b7370","Type":"ContainerDied","Data":"ab5a8118f63c4c7931595ccce783ae3eadb46bc15db93a51cedd400b130b1fd0"} Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.711185 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57d6" event={"ID":"f28a7275-4712-4dc9-89a3-16a77bba72d1","Type":"ContainerStarted","Data":"8dcea5cde4292cf3e3b5f0ecf7804687718f4456131e9b2d7bbf191afba8827b"} Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.712416 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" event={"ID":"d284ca9d-999e-4ab5-8d7e-647af799d69c","Type":"ContainerStarted","Data":"dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952"} Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.713829 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjxbl" event={"ID":"c9ebbd26-bcdc-4db3-b345-a5715ac59a82","Type":"ContainerStarted","Data":"ded987bb7da373d6123e13a47a53040e806b2aaf10006f0503aacf00428712ab"} Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.748243 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mnvn\" (UniqueName: \"kubernetes.io/projected/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-kube-api-access-7mnvn\") pod \"redhat-operators-vfz9w\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.748299 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-catalog-content\") pod \"redhat-operators-vfz9w\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.748384 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-utilities\") pod \"redhat-operators-vfz9w\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.749005 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-utilities\") pod \"redhat-operators-vfz9w\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.749314 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-catalog-content\") pod \"redhat-operators-vfz9w\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.767786 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mnvn\" (UniqueName: \"kubernetes.io/projected/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-kube-api-access-7mnvn\") pod \"redhat-operators-vfz9w\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.897017 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.970089 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wpj95"] Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.971631 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:27 crc kubenswrapper[4954]: I1209 16:59:27.979330 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wpj95"] Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.202306 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vvfc\" (UniqueName: \"kubernetes.io/projected/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-kube-api-access-9vvfc\") pod \"redhat-operators-wpj95\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.202410 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-catalog-content\") pod \"redhat-operators-wpj95\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.202493 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-utilities\") pod \"redhat-operators-wpj95\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.303511 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-utilities\") pod \"redhat-operators-wpj95\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.303619 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vvfc\" (UniqueName: \"kubernetes.io/projected/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-kube-api-access-9vvfc\") pod \"redhat-operators-wpj95\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.303669 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-catalog-content\") pod \"redhat-operators-wpj95\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.304632 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-utilities\") pod \"redhat-operators-wpj95\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.304663 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-catalog-content\") pod \"redhat-operators-wpj95\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.323748 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vvfc\" (UniqueName: \"kubernetes.io/projected/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-kube-api-access-9vvfc\") pod \"redhat-operators-wpj95\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.355294 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.368937 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:28 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:28 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:28 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.369012 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.465704 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vfz9w"] Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.729102 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vfz9w" event={"ID":"d0dc2999-1020-4b4b-96ae-87ba32e34ad8","Type":"ContainerStarted","Data":"80bc71a0bf53a88cfdada5e765b97516cfd9ba88ebeeef6cb451e3c7a4c8f87b"} Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.733146 4954 generic.go:334] "Generic (PLEG): container finished" podID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerID="44412eee609a047930cd88f51f8174332c91aba076269be6583ddfc271cf63ed" exitCode=0 Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.733196 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57d6" event={"ID":"f28a7275-4712-4dc9-89a3-16a77bba72d1","Type":"ContainerDied","Data":"44412eee609a047930cd88f51f8174332c91aba076269be6583ddfc271cf63ed"} Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.734808 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.734944 4954 generic.go:334] "Generic (PLEG): container finished" podID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerID="bb637abb2df47f522f090fb6b4550c185f4cc3f38e406089a64aa08526d9a168" exitCode=0 Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.734982 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjxbl" event={"ID":"c9ebbd26-bcdc-4db3-b345-a5715ac59a82","Type":"ContainerDied","Data":"bb637abb2df47f522f090fb6b4550c185f4cc3f38e406089a64aa08526d9a168"} Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.741126 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" event={"ID":"856fc314-24bd-403e-a4b0-3dcd73eba595","Type":"ContainerStarted","Data":"08e19b35553a6540dde8b822b742d57a2dae39ef53ccd1ebece979299cf629c5"} Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.743009 4954 generic.go:334] "Generic (PLEG): container finished" podID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerID="10bf1fccc4e2742455309e9f198f00a2e607aa3a3def637556a839c0e59ade8c" exitCode=0 Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.743242 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nlf5c" event={"ID":"271559e2-f374-4aeb-a3f7-6b039afc7e8f","Type":"ContainerDied","Data":"10bf1fccc4e2742455309e9f198f00a2e607aa3a3def637556a839c0e59ade8c"} Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.744674 4954 generic.go:334] "Generic (PLEG): container finished" podID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerID="3f6fd90edf925a9aae6462e32442f363cc5fbf2f72578b67612d6f3c119a7035" exitCode=0 Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.744716 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-snfrc" event={"ID":"1df4d5f7-03d7-4b91-b949-b555dae7ce2b","Type":"ContainerDied","Data":"3f6fd90edf925a9aae6462e32442f363cc5fbf2f72578b67612d6f3c119a7035"} Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.746565 4954 generic.go:334] "Generic (PLEG): container finished" podID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerID="78a8cfdf2f09063c9e057115de3ecad74edf3bb4dd5d3017be89c2c80dd6f243" exitCode=0 Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.746616 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9xz9" event={"ID":"f344130d-4bc2-4575-bb60-815ea6ae6a84","Type":"ContainerDied","Data":"78a8cfdf2f09063c9e057115de3ecad74edf3bb4dd5d3017be89c2c80dd6f243"} Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.748448 4954 generic.go:334] "Generic (PLEG): container finished" podID="bcc59fe2-e410-4646-be11-560d0f45938b" containerID="d60e56fbf6d62aa1fbc2d97a997ebb9f713fa11ff65c52000c0826f71ed9655c" exitCode=0 Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.749367 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7mwfq" event={"ID":"bcc59fe2-e410-4646-be11-560d0f45938b","Type":"ContainerDied","Data":"d60e56fbf6d62aa1fbc2d97a997ebb9f713fa11ff65c52000c0826f71ed9655c"} Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.749388 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:28 crc kubenswrapper[4954]: I1209 16:59:28.994494 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" podStartSLOduration=146.994469482 podStartE2EDuration="2m26.994469482s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:28.951491727 +0000 UTC m=+165.339665567" watchObservedRunningTime="2025-12-09 16:59:28.994469482 +0000 UTC m=+165.382643302" Dec 09 16:59:29 crc kubenswrapper[4954]: I1209 16:59:29.126270 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wpj95"] Dec 09 16:59:29 crc kubenswrapper[4954]: I1209 16:59:29.376143 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:29 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:29 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:29 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:29 crc kubenswrapper[4954]: I1209 16:59:29.376198 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:29 crc kubenswrapper[4954]: I1209 16:59:29.525493 4954 patch_prober.go:28] interesting pod/apiserver-76f77b778f-v6ftv container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]log ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]etcd ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]poststarthook/generic-apiserver-start-informers ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]poststarthook/max-in-flight-filter ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 09 16:59:29 crc kubenswrapper[4954]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 09 16:59:29 crc kubenswrapper[4954]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 09 16:59:29 crc kubenswrapper[4954]: [+]poststarthook/project.openshift.io-projectcache ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]poststarthook/openshift.io-startinformers ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 09 16:59:29 crc kubenswrapper[4954]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 09 16:59:29 crc kubenswrapper[4954]: livez check failed Dec 09 16:59:29 crc kubenswrapper[4954]: I1209 16:59:29.525559 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" podUID="e43b0681-33e3-4eca-a737-ed3c9c2fe994" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:29 crc kubenswrapper[4954]: I1209 16:59:29.755660 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpj95" event={"ID":"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56","Type":"ContainerStarted","Data":"cc5f1f84b1d876464d860fe9634e3634e3c15a18001e682c9fef52b1ee4970be"} Dec 09 16:59:29 crc kubenswrapper[4954]: I1209 16:59:29.757560 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"981f3a4c-bcf3-4a51-af62-28a3f65b7370","Type":"ContainerDied","Data":"ac50edf9af0ce63241997d865caa3fdf2fa64b84b2b030d8697a533d36a0dccb"} Dec 09 16:59:29 crc kubenswrapper[4954]: I1209 16:59:29.757767 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac50edf9af0ce63241997d865caa3fdf2fa64b84b2b030d8697a533d36a0dccb" Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.059195 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.165050 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kubelet-dir\") pod \"981f3a4c-bcf3-4a51-af62-28a3f65b7370\" (UID: \"981f3a4c-bcf3-4a51-af62-28a3f65b7370\") " Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.165179 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kube-api-access\") pod \"981f3a4c-bcf3-4a51-af62-28a3f65b7370\" (UID: \"981f3a4c-bcf3-4a51-af62-28a3f65b7370\") " Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.165175 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "981f3a4c-bcf3-4a51-af62-28a3f65b7370" (UID: "981f3a4c-bcf3-4a51-af62-28a3f65b7370"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.165478 4954 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.238979 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "981f3a4c-bcf3-4a51-af62-28a3f65b7370" (UID: "981f3a4c-bcf3-4a51-af62-28a3f65b7370"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.266729 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/981f3a4c-bcf3-4a51-af62-28a3f65b7370-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.368792 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:30 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:30 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:30 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.368888 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.393431 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.393529 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.393805 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.393881 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.615000 4954 patch_prober.go:28] interesting pod/console-f9d7485db-v4pm7 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.39:8443/health\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 09 16:59:30 crc kubenswrapper[4954]: I1209 16:59:30.615401 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-v4pm7" podUID="7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.39:8443/health\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.069666 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ffkzn" event={"ID":"856fc314-24bd-403e-a4b0-3dcd73eba595","Type":"ContainerStarted","Data":"58d0e0bc6676d96fa79607c4518764f8c8a4f342d043646816f53b509892c279"} Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.076970 4954 generic.go:334] "Generic (PLEG): container finished" podID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerID="32a2781665742e8ec23731658a2f67bd96dcd24604b3a613a3ce82a2d575a87e" exitCode=0 Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.077072 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpj95" event={"ID":"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56","Type":"ContainerDied","Data":"32a2781665742e8ec23731658a2f67bd96dcd24604b3a613a3ce82a2d575a87e"} Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.088053 4954 generic.go:334] "Generic (PLEG): container finished" podID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerID="7b11d65a4c6d573ca717ed9eddb0f0fcaa8c176aa01652ee0d728205f85892e4" exitCode=0 Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.088417 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vfz9w" event={"ID":"d0dc2999-1020-4b4b-96ae-87ba32e34ad8","Type":"ContainerDied","Data":"7b11d65a4c6d573ca717ed9eddb0f0fcaa8c176aa01652ee0d728205f85892e4"} Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.088676 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.097330 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.097656 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-ffkzn" podStartSLOduration=149.097646109 podStartE2EDuration="2m29.097646109s" podCreationTimestamp="2025-12-09 16:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:59:31.088282824 +0000 UTC m=+167.476456644" watchObservedRunningTime="2025-12-09 16:59:31.097646109 +0000 UTC m=+167.485819929" Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.101972 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-v6ftv" Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.412497 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:31 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:31 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:31 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.412540 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.416885 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 16:59:31 crc kubenswrapper[4954]: I1209 16:59:31.549806 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pp8mx" Dec 09 16:59:32 crc kubenswrapper[4954]: I1209 16:59:32.383370 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:32 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:32 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:32 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:32 crc kubenswrapper[4954]: I1209 16:59:32.383428 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:33 crc kubenswrapper[4954]: I1209 16:59:33.471246 4954 patch_prober.go:28] interesting pod/router-default-5444994796-cj74l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 16:59:33 crc kubenswrapper[4954]: [-]has-synced failed: reason withheld Dec 09 16:59:33 crc kubenswrapper[4954]: [+]process-running ok Dec 09 16:59:33 crc kubenswrapper[4954]: healthz check failed Dec 09 16:59:33 crc kubenswrapper[4954]: I1209 16:59:33.471300 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cj74l" podUID="cbebf2b4-7952-45d6-8c77-efc6438e05fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 16:59:34 crc kubenswrapper[4954]: I1209 16:59:34.376153 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:34 crc kubenswrapper[4954]: I1209 16:59:34.381581 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-cj74l" Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.389407 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.389485 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.389880 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.389942 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.390008 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.390535 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.390587 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.391099 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"a49768838bf2298ccc0abf590b58a6ff051c216609e208f0dfdeead43339ee1a"} pod="openshift-console/downloads-7954f5f757-6gmvm" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.391220 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" containerID="cri-o://a49768838bf2298ccc0abf590b58a6ff051c216609e208f0dfdeead43339ee1a" gracePeriod=2 Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.620567 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.624469 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.729827 4954 generic.go:334] "Generic (PLEG): container finished" podID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerID="a49768838bf2298ccc0abf590b58a6ff051c216609e208f0dfdeead43339ee1a" exitCode=0 Dec 09 16:59:40 crc kubenswrapper[4954]: I1209 16:59:40.729875 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6gmvm" event={"ID":"dc11e3bb-97da-485b-ad28-34b0116ed220","Type":"ContainerDied","Data":"a49768838bf2298ccc0abf590b58a6ff051c216609e208f0dfdeead43339ee1a"} Dec 09 16:59:42 crc kubenswrapper[4954]: I1209 16:59:42.777355 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6gmvm" event={"ID":"dc11e3bb-97da-485b-ad28-34b0116ed220","Type":"ContainerStarted","Data":"7896b1f320892fa8a18675230e1a1853db1cac11c5d5db41b6ec51dfddaebd80"} Dec 09 16:59:43 crc kubenswrapper[4954]: I1209 16:59:43.379820 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 16:59:43 crc kubenswrapper[4954]: I1209 16:59:43.772678 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:59:43 crc kubenswrapper[4954]: I1209 16:59:43.772763 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:59:43 crc kubenswrapper[4954]: I1209 16:59:43.792415 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 16:59:43 crc kubenswrapper[4954]: I1209 16:59:43.793790 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:43 crc kubenswrapper[4954]: I1209 16:59:43.793869 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:44 crc kubenswrapper[4954]: I1209 16:59:44.798475 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:44 crc kubenswrapper[4954]: I1209 16:59:44.798842 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:50 crc kubenswrapper[4954]: I1209 16:59:50.520269 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:50 crc kubenswrapper[4954]: I1209 16:59:50.520338 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 16:59:50 crc kubenswrapper[4954]: I1209 16:59:50.520825 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:50 crc kubenswrapper[4954]: I1209 16:59:50.520709 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 16:59:51 crc kubenswrapper[4954]: I1209 16:59:51.525836 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 16:59:51 crc kubenswrapper[4954]: I1209 16:59:51.556227 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cbfp9" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.522207 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 16:59:57 crc kubenswrapper[4954]: E1209 16:59:57.523382 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="981f3a4c-bcf3-4a51-af62-28a3f65b7370" containerName="pruner" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.524265 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="981f3a4c-bcf3-4a51-af62-28a3f65b7370" containerName="pruner" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.524857 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="981f3a4c-bcf3-4a51-af62-28a3f65b7370" containerName="pruner" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.531398 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.534628 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.534685 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.545187 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.721879 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e681c023-d702-4592-870d-0e8253c077a5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e681c023-d702-4592-870d-0e8253c077a5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.722761 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e681c023-d702-4592-870d-0e8253c077a5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e681c023-d702-4592-870d-0e8253c077a5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.823138 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e681c023-d702-4592-870d-0e8253c077a5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e681c023-d702-4592-870d-0e8253c077a5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.823186 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e681c023-d702-4592-870d-0e8253c077a5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e681c023-d702-4592-870d-0e8253c077a5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.823380 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e681c023-d702-4592-870d-0e8253c077a5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e681c023-d702-4592-870d-0e8253c077a5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.840520 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e681c023-d702-4592-870d-0e8253c077a5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e681c023-d702-4592-870d-0e8253c077a5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 16:59:57 crc kubenswrapper[4954]: I1209 16:59:57.863039 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.134145 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd"] Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.135626 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.139510 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.139835 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.144206 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd"] Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.269964 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-secret-volume\") pod \"collect-profiles-29421660-fnwcd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.270019 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-config-volume\") pod \"collect-profiles-29421660-fnwcd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.270059 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7klq\" (UniqueName: \"kubernetes.io/projected/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-kube-api-access-s7klq\") pod \"collect-profiles-29421660-fnwcd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.371000 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-secret-volume\") pod \"collect-profiles-29421660-fnwcd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.371056 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-config-volume\") pod \"collect-profiles-29421660-fnwcd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.371088 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7klq\" (UniqueName: \"kubernetes.io/projected/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-kube-api-access-s7klq\") pod \"collect-profiles-29421660-fnwcd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.372304 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-config-volume\") pod \"collect-profiles-29421660-fnwcd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.374822 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-secret-volume\") pod \"collect-profiles-29421660-fnwcd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.387578 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7klq\" (UniqueName: \"kubernetes.io/projected/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-kube-api-access-s7klq\") pod \"collect-profiles-29421660-fnwcd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.389692 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.389731 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.389763 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.389843 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 17:00:00 crc kubenswrapper[4954]: I1209 17:00:00.467114 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.315408 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.316103 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.327031 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.404824 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-var-lock\") pod \"installer-9-crc\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.404957 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kube-api-access\") pod \"installer-9-crc\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.405127 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.506759 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.506839 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-var-lock\") pod \"installer-9-crc\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.506860 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kube-api-access\") pod \"installer-9-crc\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.507177 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.507214 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-var-lock\") pod \"installer-9-crc\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.523850 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kube-api-access\") pod \"installer-9-crc\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:02 crc kubenswrapper[4954]: I1209 17:00:02.632440 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:00:10 crc kubenswrapper[4954]: I1209 17:00:10.390025 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 17:00:10 crc kubenswrapper[4954]: I1209 17:00:10.390616 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 17:00:10 crc kubenswrapper[4954]: I1209 17:00:10.390145 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 17:00:10 crc kubenswrapper[4954]: I1209 17:00:10.390713 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 17:00:10 crc kubenswrapper[4954]: I1209 17:00:10.390759 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 17:00:10 crc kubenswrapper[4954]: I1209 17:00:10.391313 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"7896b1f320892fa8a18675230e1a1853db1cac11c5d5db41b6ec51dfddaebd80"} pod="openshift-console/downloads-7954f5f757-6gmvm" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 09 17:00:10 crc kubenswrapper[4954]: I1209 17:00:10.391341 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" containerID="cri-o://7896b1f320892fa8a18675230e1a1853db1cac11c5d5db41b6ec51dfddaebd80" gracePeriod=2 Dec 09 17:00:10 crc kubenswrapper[4954]: I1209 17:00:10.391408 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 17:00:10 crc kubenswrapper[4954]: I1209 17:00:10.391438 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 17:00:13 crc kubenswrapper[4954]: I1209 17:00:13.289639 4954 generic.go:334] "Generic (PLEG): container finished" podID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerID="7896b1f320892fa8a18675230e1a1853db1cac11c5d5db41b6ec51dfddaebd80" exitCode=0 Dec 09 17:00:13 crc kubenswrapper[4954]: I1209 17:00:13.289636 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6gmvm" event={"ID":"dc11e3bb-97da-485b-ad28-34b0116ed220","Type":"ContainerDied","Data":"7896b1f320892fa8a18675230e1a1853db1cac11c5d5db41b6ec51dfddaebd80"} Dec 09 17:00:13 crc kubenswrapper[4954]: I1209 17:00:13.290036 4954 scope.go:117] "RemoveContainer" containerID="a49768838bf2298ccc0abf590b58a6ff051c216609e208f0dfdeead43339ee1a" Dec 09 17:00:13 crc kubenswrapper[4954]: I1209 17:00:13.755942 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:00:13 crc kubenswrapper[4954]: I1209 17:00:13.756010 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:00:13 crc kubenswrapper[4954]: I1209 17:00:13.756059 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:00:13 crc kubenswrapper[4954]: I1209 17:00:13.757100 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:00:13 crc kubenswrapper[4954]: I1209 17:00:13.757197 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2" gracePeriod=600 Dec 09 17:00:20 crc kubenswrapper[4954]: I1209 17:00:20.326896 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2" exitCode=0 Dec 09 17:00:20 crc kubenswrapper[4954]: I1209 17:00:20.327007 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2"} Dec 09 17:00:20 crc kubenswrapper[4954]: I1209 17:00:20.389842 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 17:00:20 crc kubenswrapper[4954]: I1209 17:00:20.389908 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 17:00:23 crc kubenswrapper[4954]: E1209 17:00:23.539853 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 09 17:00:23 crc kubenswrapper[4954]: E1209 17:00:23.545579 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qlcll,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-jjxbl_openshift-marketplace(c9ebbd26-bcdc-4db3-b345-a5715ac59a82): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:00:23 crc kubenswrapper[4954]: E1209 17:00:23.546977 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-jjxbl" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" Dec 09 17:00:24 crc kubenswrapper[4954]: E1209 17:00:24.532492 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-jjxbl" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" Dec 09 17:00:24 crc kubenswrapper[4954]: E1209 17:00:24.594542 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 09 17:00:24 crc kubenswrapper[4954]: E1209 17:00:24.594710 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sjj6v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7mwfq_openshift-marketplace(bcc59fe2-e410-4646-be11-560d0f45938b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:00:24 crc kubenswrapper[4954]: E1209 17:00:24.596025 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-7mwfq" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" Dec 09 17:00:29 crc kubenswrapper[4954]: E1209 17:00:29.632695 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7mwfq" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" Dec 09 17:00:29 crc kubenswrapper[4954]: E1209 17:00:29.712999 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 09 17:00:29 crc kubenswrapper[4954]: E1209 17:00:29.713139 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9vvfc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-wpj95_openshift-marketplace(a5e4dfeb-181d-43c7-89bb-3c828fb5ce56): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:00:29 crc kubenswrapper[4954]: E1209 17:00:29.715675 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-wpj95" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" Dec 09 17:00:30 crc kubenswrapper[4954]: I1209 17:00:30.389998 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 17:00:30 crc kubenswrapper[4954]: I1209 17:00:30.390494 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.080612 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-wpj95" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.168950 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.169113 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2zdx7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-snfrc_openshift-marketplace(1df4d5f7-03d7-4b91-b949-b555dae7ce2b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.170270 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-snfrc" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.191746 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.191895 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9pzhr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-nlf5c_openshift-marketplace(271559e2-f374-4aeb-a3f7-6b039afc7e8f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.193819 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-nlf5c" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.212716 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.213072 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sjprj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-c9xz9_openshift-marketplace(f344130d-4bc2-4575-bb60-815ea6ae6a84): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.214294 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-c9xz9" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.248631 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.248807 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7mnvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-vfz9w_openshift-marketplace(d0dc2999-1020-4b4b-96ae-87ba32e34ad8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.250217 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-vfz9w" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.264159 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.264292 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nmdfn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-j57d6_openshift-marketplace(f28a7275-4712-4dc9-89a3-16a77bba72d1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.266334 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-j57d6" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" Dec 09 17:00:31 crc kubenswrapper[4954]: I1209 17:00:31.408074 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6gmvm" event={"ID":"dc11e3bb-97da-485b-ad28-34b0116ed220","Type":"ContainerStarted","Data":"e039d18cdb286bb4529f0f0b6f2e82f83ef7c2fbf59b64e89020d4dfc619a7d5"} Dec 09 17:00:31 crc kubenswrapper[4954]: I1209 17:00:31.408681 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 17:00:31 crc kubenswrapper[4954]: I1209 17:00:31.408730 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 17:00:31 crc kubenswrapper[4954]: I1209 17:00:31.408758 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 17:00:31 crc kubenswrapper[4954]: I1209 17:00:31.412004 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"33a69fad0fc5c3316079e821a4a615a1b8c30db3b75d7d52d707b8d7b04f4a32"} Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.627547 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-nlf5c" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.627552 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-vfz9w" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.627618 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-snfrc" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.628463 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-c9xz9" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" Dec 09 17:00:31 crc kubenswrapper[4954]: E1209 17:00:31.629158 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-j57d6" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" Dec 09 17:00:31 crc kubenswrapper[4954]: I1209 17:00:31.635515 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd"] Dec 09 17:00:31 crc kubenswrapper[4954]: I1209 17:00:31.640374 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 17:00:31 crc kubenswrapper[4954]: W1209 17:00:31.652663 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfb7d6bd_431a_4dc7_a2e3_d3d4850deabd.slice/crio-3e854ef941287d3839638d2b6b32d530cc62b9b92c6d598e4026a4655db1e713 WatchSource:0}: Error finding container 3e854ef941287d3839638d2b6b32d530cc62b9b92c6d598e4026a4655db1e713: Status 404 returned error can't find the container with id 3e854ef941287d3839638d2b6b32d530cc62b9b92c6d598e4026a4655db1e713 Dec 09 17:00:31 crc kubenswrapper[4954]: W1209 17:00:31.658059 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode681c023_d702_4592_870d_0e8253c077a5.slice/crio-c309c383cd1dc0c2963a2c6a2b2ee29c2308fbe37bdc45abd643536b426f9c45 WatchSource:0}: Error finding container c309c383cd1dc0c2963a2c6a2b2ee29c2308fbe37bdc45abd643536b426f9c45: Status 404 returned error can't find the container with id c309c383cd1dc0c2963a2c6a2b2ee29c2308fbe37bdc45abd643536b426f9c45 Dec 09 17:00:31 crc kubenswrapper[4954]: I1209 17:00:31.717489 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.418880 4954 patch_prober.go:28] interesting pod/downloads-7954f5f757-6gmvm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.419160 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6gmvm" podUID="dc11e3bb-97da-485b-ad28-34b0116ed220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.421082 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e681c023-d702-4592-870d-0e8253c077a5","Type":"ContainerStarted","Data":"55d08620c3d2cf1f22a118c25c3c5260c5fe8a73c6cd628696194b2fbd8a298a"} Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.421155 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e681c023-d702-4592-870d-0e8253c077a5","Type":"ContainerStarted","Data":"c309c383cd1dc0c2963a2c6a2b2ee29c2308fbe37bdc45abd643536b426f9c45"} Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.423979 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ce53dd1a-0beb-4c91-9422-8473bd7a9412","Type":"ContainerStarted","Data":"518372519c39eeb1f94b18331356f455b2c61dfa71a9fa822d1fa67c6cc7e937"} Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.424022 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ce53dd1a-0beb-4c91-9422-8473bd7a9412","Type":"ContainerStarted","Data":"255c516d4832905ae7775e3be235f197ad3d6a0cc3957dd56c6810174fde4db7"} Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.427985 4954 generic.go:334] "Generic (PLEG): container finished" podID="dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd" containerID="00fff6ceec063b5787c807e6e60ca4a38ec961ab1ea3c8d36d3cb655f87502e1" exitCode=0 Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.428025 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" event={"ID":"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd","Type":"ContainerDied","Data":"00fff6ceec063b5787c807e6e60ca4a38ec961ab1ea3c8d36d3cb655f87502e1"} Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.428055 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" event={"ID":"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd","Type":"ContainerStarted","Data":"3e854ef941287d3839638d2b6b32d530cc62b9b92c6d598e4026a4655db1e713"} Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.439120 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=35.43909867 podStartE2EDuration="35.43909867s" podCreationTimestamp="2025-12-09 16:59:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:00:32.435455846 +0000 UTC m=+228.823629666" watchObservedRunningTime="2025-12-09 17:00:32.43909867 +0000 UTC m=+228.827272490" Dec 09 17:00:32 crc kubenswrapper[4954]: I1209 17:00:32.466187 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=30.466169864 podStartE2EDuration="30.466169864s" podCreationTimestamp="2025-12-09 17:00:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:00:32.463770639 +0000 UTC m=+228.851944469" watchObservedRunningTime="2025-12-09 17:00:32.466169864 +0000 UTC m=+228.854343674" Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.182181 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6449l"] Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.443968 4954 generic.go:334] "Generic (PLEG): container finished" podID="e681c023-d702-4592-870d-0e8253c077a5" containerID="55d08620c3d2cf1f22a118c25c3c5260c5fe8a73c6cd628696194b2fbd8a298a" exitCode=0 Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.444113 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e681c023-d702-4592-870d-0e8253c077a5","Type":"ContainerDied","Data":"55d08620c3d2cf1f22a118c25c3c5260c5fe8a73c6cd628696194b2fbd8a298a"} Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.740730 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.801081 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-secret-volume\") pod \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.801204 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-config-volume\") pod \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.801276 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7klq\" (UniqueName: \"kubernetes.io/projected/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-kube-api-access-s7klq\") pod \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\" (UID: \"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd\") " Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.802411 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-config-volume" (OuterVolumeSpecName: "config-volume") pod "dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd" (UID: "dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.809352 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd" (UID: "dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.813087 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-kube-api-access-s7klq" (OuterVolumeSpecName: "kube-api-access-s7klq") pod "dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd" (UID: "dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd"). InnerVolumeSpecName "kube-api-access-s7klq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.903087 4954 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.903149 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7klq\" (UniqueName: \"kubernetes.io/projected/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-kube-api-access-s7klq\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:33 crc kubenswrapper[4954]: I1209 17:00:33.903168 4954 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:34 crc kubenswrapper[4954]: I1209 17:00:34.449876 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" Dec 09 17:00:34 crc kubenswrapper[4954]: I1209 17:00:34.450362 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd" event={"ID":"dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd","Type":"ContainerDied","Data":"3e854ef941287d3839638d2b6b32d530cc62b9b92c6d598e4026a4655db1e713"} Dec 09 17:00:34 crc kubenswrapper[4954]: I1209 17:00:34.450388 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e854ef941287d3839638d2b6b32d530cc62b9b92c6d598e4026a4655db1e713" Dec 09 17:00:34 crc kubenswrapper[4954]: I1209 17:00:34.678098 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 17:00:34 crc kubenswrapper[4954]: I1209 17:00:34.739546 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e681c023-d702-4592-870d-0e8253c077a5-kube-api-access\") pod \"e681c023-d702-4592-870d-0e8253c077a5\" (UID: \"e681c023-d702-4592-870d-0e8253c077a5\") " Dec 09 17:00:34 crc kubenswrapper[4954]: I1209 17:00:34.739709 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e681c023-d702-4592-870d-0e8253c077a5-kubelet-dir\") pod \"e681c023-d702-4592-870d-0e8253c077a5\" (UID: \"e681c023-d702-4592-870d-0e8253c077a5\") " Dec 09 17:00:34 crc kubenswrapper[4954]: I1209 17:00:34.739900 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e681c023-d702-4592-870d-0e8253c077a5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e681c023-d702-4592-870d-0e8253c077a5" (UID: "e681c023-d702-4592-870d-0e8253c077a5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:00:34 crc kubenswrapper[4954]: I1209 17:00:34.743787 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e681c023-d702-4592-870d-0e8253c077a5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e681c023-d702-4592-870d-0e8253c077a5" (UID: "e681c023-d702-4592-870d-0e8253c077a5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:00:34 crc kubenswrapper[4954]: I1209 17:00:34.841245 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e681c023-d702-4592-870d-0e8253c077a5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:34 crc kubenswrapper[4954]: I1209 17:00:34.841287 4954 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e681c023-d702-4592-870d-0e8253c077a5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:35 crc kubenswrapper[4954]: I1209 17:00:35.455637 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e681c023-d702-4592-870d-0e8253c077a5","Type":"ContainerDied","Data":"c309c383cd1dc0c2963a2c6a2b2ee29c2308fbe37bdc45abd643536b426f9c45"} Dec 09 17:00:35 crc kubenswrapper[4954]: I1209 17:00:35.455685 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c309c383cd1dc0c2963a2c6a2b2ee29c2308fbe37bdc45abd643536b426f9c45" Dec 09 17:00:35 crc kubenswrapper[4954]: I1209 17:00:35.455659 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 17:00:40 crc kubenswrapper[4954]: I1209 17:00:40.396531 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6gmvm" Dec 09 17:00:42 crc kubenswrapper[4954]: I1209 17:00:42.502439 4954 generic.go:334] "Generic (PLEG): container finished" podID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerID="cf55d9085bdaf7b9669d4fd8b2b8146e506b22cc4c443935b95201e5c900f380" exitCode=0 Dec 09 17:00:42 crc kubenswrapper[4954]: I1209 17:00:42.502531 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjxbl" event={"ID":"c9ebbd26-bcdc-4db3-b345-a5715ac59a82","Type":"ContainerDied","Data":"cf55d9085bdaf7b9669d4fd8b2b8146e506b22cc4c443935b95201e5c900f380"} Dec 09 17:00:43 crc kubenswrapper[4954]: I1209 17:00:43.509912 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nlf5c" event={"ID":"271559e2-f374-4aeb-a3f7-6b039afc7e8f","Type":"ContainerStarted","Data":"968568e4eee1faa4569a82635ca2e523923a4b9f547c7642c467122a91ef6cf1"} Dec 09 17:00:44 crc kubenswrapper[4954]: I1209 17:00:44.520046 4954 generic.go:334] "Generic (PLEG): container finished" podID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerID="968568e4eee1faa4569a82635ca2e523923a4b9f547c7642c467122a91ef6cf1" exitCode=0 Dec 09 17:00:44 crc kubenswrapper[4954]: I1209 17:00:44.520132 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nlf5c" event={"ID":"271559e2-f374-4aeb-a3f7-6b039afc7e8f","Type":"ContainerDied","Data":"968568e4eee1faa4569a82635ca2e523923a4b9f547c7642c467122a91ef6cf1"} Dec 09 17:00:44 crc kubenswrapper[4954]: I1209 17:00:44.522559 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vfz9w" event={"ID":"d0dc2999-1020-4b4b-96ae-87ba32e34ad8","Type":"ContainerStarted","Data":"b25bdc1a0ac311120b1b070d863514e204cf6251afe24724215fe7a7dff45d43"} Dec 09 17:00:45 crc kubenswrapper[4954]: I1209 17:00:45.529505 4954 generic.go:334] "Generic (PLEG): container finished" podID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerID="b25bdc1a0ac311120b1b070d863514e204cf6251afe24724215fe7a7dff45d43" exitCode=0 Dec 09 17:00:45 crc kubenswrapper[4954]: I1209 17:00:45.529563 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vfz9w" event={"ID":"d0dc2999-1020-4b4b-96ae-87ba32e34ad8","Type":"ContainerDied","Data":"b25bdc1a0ac311120b1b070d863514e204cf6251afe24724215fe7a7dff45d43"} Dec 09 17:00:58 crc kubenswrapper[4954]: I1209 17:00:58.212320 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" podUID="9ada019a-2740-482e-864b-55762c330b26" containerName="oauth-openshift" containerID="cri-o://de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c" gracePeriod=15 Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.454506 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.486029 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w"] Dec 09 17:00:59 crc kubenswrapper[4954]: E1209 17:00:59.486238 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ada019a-2740-482e-864b-55762c330b26" containerName="oauth-openshift" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.486251 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ada019a-2740-482e-864b-55762c330b26" containerName="oauth-openshift" Dec 09 17:00:59 crc kubenswrapper[4954]: E1209 17:00:59.486260 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e681c023-d702-4592-870d-0e8253c077a5" containerName="pruner" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.486267 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e681c023-d702-4592-870d-0e8253c077a5" containerName="pruner" Dec 09 17:00:59 crc kubenswrapper[4954]: E1209 17:00:59.486278 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd" containerName="collect-profiles" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.486286 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd" containerName="collect-profiles" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.486379 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd" containerName="collect-profiles" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.486390 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ada019a-2740-482e-864b-55762c330b26" containerName="oauth-openshift" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.486406 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="e681c023-d702-4592-870d-0e8253c077a5" containerName="pruner" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.486795 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.498304 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w"] Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508679 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-session\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508717 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-service-ca\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508743 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-login\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508759 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-audit-policies\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508787 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-router-certs\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508836 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-serving-cert\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508860 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-ocp-branding-template\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508882 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-error\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508929 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-cliconfig\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508953 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-idp-0-file-data\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.508974 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-provider-selection\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.509002 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhkx9\" (UniqueName: \"kubernetes.io/projected/9ada019a-2740-482e-864b-55762c330b26-kube-api-access-rhkx9\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.509029 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9ada019a-2740-482e-864b-55762c330b26-audit-dir\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.509076 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-trusted-ca-bundle\") pod \"9ada019a-2740-482e-864b-55762c330b26\" (UID: \"9ada019a-2740-482e-864b-55762c330b26\") " Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.510171 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.532151 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.532914 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.533543 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.534437 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ada019a-2740-482e-864b-55762c330b26-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.536736 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.537383 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.538458 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.540988 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.541888 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ada019a-2740-482e-864b-55762c330b26-kube-api-access-rhkx9" (OuterVolumeSpecName: "kube-api-access-rhkx9") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "kube-api-access-rhkx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.544749 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.544985 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.545160 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.545430 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "9ada019a-2740-482e-864b-55762c330b26" (UID: "9ada019a-2740-482e-864b-55762c330b26"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.610846 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-audit-policies\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.610908 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-session\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.610939 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/896786b7-37f5-421d-a3eb-b5393d826bbf-audit-dir\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.610966 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-router-certs\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611021 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qt4h\" (UniqueName: \"kubernetes.io/projected/896786b7-37f5-421d-a3eb-b5393d826bbf-kube-api-access-7qt4h\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611043 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-template-login\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611073 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611140 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611179 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-template-error\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611198 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611223 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611243 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611269 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-service-ca\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611285 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611515 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611552 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611577 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611635 4954 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611657 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611679 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611700 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611720 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611739 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611757 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611775 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611795 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhkx9\" (UniqueName: \"kubernetes.io/projected/9ada019a-2740-482e-864b-55762c330b26-kube-api-access-rhkx9\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611812 4954 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9ada019a-2740-482e-864b-55762c330b26-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.611829 4954 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ada019a-2740-482e-864b-55762c330b26-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.617831 4954 generic.go:334] "Generic (PLEG): container finished" podID="9ada019a-2740-482e-864b-55762c330b26" containerID="de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c" exitCode=0 Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.617894 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.617891 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" event={"ID":"9ada019a-2740-482e-864b-55762c330b26","Type":"ContainerDied","Data":"de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c"} Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.618130 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6449l" event={"ID":"9ada019a-2740-482e-864b-55762c330b26","Type":"ContainerDied","Data":"5537b743e782b6973efa1383bf6201177d90e5a128bcb21306a04919ec7ed18f"} Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.618179 4954 scope.go:117] "RemoveContainer" containerID="de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.649785 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6449l"] Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.653615 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6449l"] Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713187 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/896786b7-37f5-421d-a3eb-b5393d826bbf-audit-dir\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713250 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-router-certs\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713295 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qt4h\" (UniqueName: \"kubernetes.io/projected/896786b7-37f5-421d-a3eb-b5393d826bbf-kube-api-access-7qt4h\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713319 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-template-login\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713350 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713384 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713428 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-template-error\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713451 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713475 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713496 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713524 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-service-ca\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713547 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713581 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-audit-policies\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713624 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-session\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.713330 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/896786b7-37f5-421d-a3eb-b5393d826bbf-audit-dir\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.714863 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.715079 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-service-ca\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.715258 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.715479 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/896786b7-37f5-421d-a3eb-b5393d826bbf-audit-policies\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.723703 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-template-login\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.728062 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.730133 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.730236 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-router-certs\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.731115 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-session\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.733694 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-user-template-error\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.734115 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.734659 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/896786b7-37f5-421d-a3eb-b5393d826bbf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.750785 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qt4h\" (UniqueName: \"kubernetes.io/projected/896786b7-37f5-421d-a3eb-b5393d826bbf-kube-api-access-7qt4h\") pod \"oauth-openshift-9bc7b6b6b-rwz2w\" (UID: \"896786b7-37f5-421d-a3eb-b5393d826bbf\") " pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:00:59 crc kubenswrapper[4954]: I1209 17:00:59.827769 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:01:00 crc kubenswrapper[4954]: I1209 17:01:00.127242 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ada019a-2740-482e-864b-55762c330b26" path="/var/lib/kubelet/pods/9ada019a-2740-482e-864b-55762c330b26/volumes" Dec 09 17:01:01 crc kubenswrapper[4954]: I1209 17:01:01.895149 4954 scope.go:117] "RemoveContainer" containerID="de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c" Dec 09 17:01:01 crc kubenswrapper[4954]: E1209 17:01:01.896263 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c\": container with ID starting with de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c not found: ID does not exist" containerID="de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c" Dec 09 17:01:01 crc kubenswrapper[4954]: I1209 17:01:01.896341 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c"} err="failed to get container status \"de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c\": rpc error: code = NotFound desc = could not find container \"de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c\": container with ID starting with de352b4bfb52b9e63de65d953474cedf5ba1f065516a212c04255dfd9d289f5c not found: ID does not exist" Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.391357 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w"] Dec 09 17:01:02 crc kubenswrapper[4954]: W1209 17:01:02.399860 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod896786b7_37f5_421d_a3eb_b5393d826bbf.slice/crio-1e3e0aff8ae6814f0b90cb8cf9861971dc1a7a8ff592ce13ead1ad55aceac133 WatchSource:0}: Error finding container 1e3e0aff8ae6814f0b90cb8cf9861971dc1a7a8ff592ce13ead1ad55aceac133: Status 404 returned error can't find the container with id 1e3e0aff8ae6814f0b90cb8cf9861971dc1a7a8ff592ce13ead1ad55aceac133 Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.669966 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-snfrc" event={"ID":"1df4d5f7-03d7-4b91-b949-b555dae7ce2b","Type":"ContainerStarted","Data":"4dfa72dd47b8b7a5c5f2643b19ff9e2b51912bf2e84293a6f40709a76ee57a9e"} Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.672818 4954 generic.go:334] "Generic (PLEG): container finished" podID="bcc59fe2-e410-4646-be11-560d0f45938b" containerID="31768f2a27b179b400d6e8a37d899d36900883a9451a69d53d4ff72cee5307bf" exitCode=0 Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.672919 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7mwfq" event={"ID":"bcc59fe2-e410-4646-be11-560d0f45938b","Type":"ContainerDied","Data":"31768f2a27b179b400d6e8a37d899d36900883a9451a69d53d4ff72cee5307bf"} Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.676054 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vfz9w" event={"ID":"d0dc2999-1020-4b4b-96ae-87ba32e34ad8","Type":"ContainerStarted","Data":"d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8"} Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.678028 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" event={"ID":"896786b7-37f5-421d-a3eb-b5393d826bbf","Type":"ContainerStarted","Data":"1e3e0aff8ae6814f0b90cb8cf9861971dc1a7a8ff592ce13ead1ad55aceac133"} Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.682195 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57d6" event={"ID":"f28a7275-4712-4dc9-89a3-16a77bba72d1","Type":"ContainerStarted","Data":"e0446210a7a4e5f9689cf10fb03e60e90f775ea68962a4cff3896e743ad926f0"} Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.684806 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjxbl" event={"ID":"c9ebbd26-bcdc-4db3-b345-a5715ac59a82","Type":"ContainerStarted","Data":"b67d954d432fdaefd19cddd78c8dc54206aa1618d67cb9535fc075be133e30b2"} Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.687619 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nlf5c" event={"ID":"271559e2-f374-4aeb-a3f7-6b039afc7e8f","Type":"ContainerStarted","Data":"76b76e8114b08f3a9678e802e06fa776197be9e1d949be861a4bbe1b9688c346"} Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.689626 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9xz9" event={"ID":"f344130d-4bc2-4575-bb60-815ea6ae6a84","Type":"ContainerStarted","Data":"7cb820c86004e8b47d2826694a262ee6143c5f7066ee1facb495127620039a5b"} Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.781345 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jjxbl" podStartSLOduration=5.566028946 podStartE2EDuration="1m38.781321815s" podCreationTimestamp="2025-12-09 16:59:24 +0000 UTC" firstStartedPulling="2025-12-09 16:59:28.737892385 +0000 UTC m=+165.126066205" lastFinishedPulling="2025-12-09 17:01:01.953185234 +0000 UTC m=+258.341359074" observedRunningTime="2025-12-09 17:01:02.77538299 +0000 UTC m=+259.163556810" watchObservedRunningTime="2025-12-09 17:01:02.781321815 +0000 UTC m=+259.169495635" Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.853306 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vfz9w" podStartSLOduration=5.007418004 podStartE2EDuration="1m35.853281867s" podCreationTimestamp="2025-12-09 16:59:27 +0000 UTC" firstStartedPulling="2025-12-09 16:59:31.096005159 +0000 UTC m=+167.484178979" lastFinishedPulling="2025-12-09 17:01:01.941869022 +0000 UTC m=+258.330042842" observedRunningTime="2025-12-09 17:01:02.85079253 +0000 UTC m=+259.238966360" watchObservedRunningTime="2025-12-09 17:01:02.853281867 +0000 UTC m=+259.241455687" Dec 09 17:01:02 crc kubenswrapper[4954]: I1209 17:01:02.873157 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nlf5c" podStartSLOduration=5.646412422 podStartE2EDuration="1m38.873133826s" podCreationTimestamp="2025-12-09 16:59:24 +0000 UTC" firstStartedPulling="2025-12-09 16:59:28.744193922 +0000 UTC m=+165.132367732" lastFinishedPulling="2025-12-09 17:01:01.970915316 +0000 UTC m=+258.359089136" observedRunningTime="2025-12-09 17:01:02.871530727 +0000 UTC m=+259.259704537" watchObservedRunningTime="2025-12-09 17:01:02.873133826 +0000 UTC m=+259.261307646" Dec 09 17:01:03 crc kubenswrapper[4954]: I1209 17:01:03.697454 4954 generic.go:334] "Generic (PLEG): container finished" podID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerID="e0446210a7a4e5f9689cf10fb03e60e90f775ea68962a4cff3896e743ad926f0" exitCode=0 Dec 09 17:01:03 crc kubenswrapper[4954]: I1209 17:01:03.697870 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57d6" event={"ID":"f28a7275-4712-4dc9-89a3-16a77bba72d1","Type":"ContainerDied","Data":"e0446210a7a4e5f9689cf10fb03e60e90f775ea68962a4cff3896e743ad926f0"} Dec 09 17:01:03 crc kubenswrapper[4954]: I1209 17:01:03.701909 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpj95" event={"ID":"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56","Type":"ContainerStarted","Data":"55a050768b735b0c202117de2c9a3f0b7f1fec8a58d5438b40597da9a6adc931"} Dec 09 17:01:03 crc kubenswrapper[4954]: I1209 17:01:03.705543 4954 generic.go:334] "Generic (PLEG): container finished" podID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerID="4dfa72dd47b8b7a5c5f2643b19ff9e2b51912bf2e84293a6f40709a76ee57a9e" exitCode=0 Dec 09 17:01:03 crc kubenswrapper[4954]: I1209 17:01:03.705615 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-snfrc" event={"ID":"1df4d5f7-03d7-4b91-b949-b555dae7ce2b","Type":"ContainerDied","Data":"4dfa72dd47b8b7a5c5f2643b19ff9e2b51912bf2e84293a6f40709a76ee57a9e"} Dec 09 17:01:03 crc kubenswrapper[4954]: I1209 17:01:03.707788 4954 generic.go:334] "Generic (PLEG): container finished" podID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerID="7cb820c86004e8b47d2826694a262ee6143c5f7066ee1facb495127620039a5b" exitCode=0 Dec 09 17:01:03 crc kubenswrapper[4954]: I1209 17:01:03.707866 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9xz9" event={"ID":"f344130d-4bc2-4575-bb60-815ea6ae6a84","Type":"ContainerDied","Data":"7cb820c86004e8b47d2826694a262ee6143c5f7066ee1facb495127620039a5b"} Dec 09 17:01:03 crc kubenswrapper[4954]: I1209 17:01:03.710671 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" event={"ID":"896786b7-37f5-421d-a3eb-b5393d826bbf","Type":"ContainerStarted","Data":"90db1e5a15b9bcc557833b62ba536252e1a8c0bfa5c6a591754bcac13193b56c"} Dec 09 17:01:03 crc kubenswrapper[4954]: I1209 17:01:03.711208 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:01:03 crc kubenswrapper[4954]: I1209 17:01:03.777480 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" podStartSLOduration=30.777457612 podStartE2EDuration="30.777457612s" podCreationTimestamp="2025-12-09 17:00:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:01:03.7732289 +0000 UTC m=+260.161402720" watchObservedRunningTime="2025-12-09 17:01:03.777457612 +0000 UTC m=+260.165631432" Dec 09 17:01:04 crc kubenswrapper[4954]: I1209 17:01:04.490673 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-9bc7b6b6b-rwz2w" Dec 09 17:01:05 crc kubenswrapper[4954]: I1209 17:01:05.190234 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 17:01:05 crc kubenswrapper[4954]: I1209 17:01:05.190300 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 17:01:05 crc kubenswrapper[4954]: I1209 17:01:05.289617 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 17:01:05 crc kubenswrapper[4954]: I1209 17:01:05.289860 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 17:01:05 crc kubenswrapper[4954]: I1209 17:01:05.820521 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-snfrc" event={"ID":"1df4d5f7-03d7-4b91-b949-b555dae7ce2b","Type":"ContainerStarted","Data":"62ba3752232989ae967c4086a4ac4d99e2ef36ed7c7dcd22835f3ba01353318b"} Dec 09 17:01:05 crc kubenswrapper[4954]: I1209 17:01:05.822658 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9xz9" event={"ID":"f344130d-4bc2-4575-bb60-815ea6ae6a84","Type":"ContainerStarted","Data":"975ac2f9c9073ce03cdb134f155cb901b644766788af76e2c3348b40c59adfed"} Dec 09 17:01:05 crc kubenswrapper[4954]: I1209 17:01:05.826024 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57d6" event={"ID":"f28a7275-4712-4dc9-89a3-16a77bba72d1","Type":"ContainerStarted","Data":"474a91cd7207e014c2d4d28ffbb23669f09afd2ee82fd6fcde0d2bfafbe7845d"} Dec 09 17:01:05 crc kubenswrapper[4954]: I1209 17:01:05.880540 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-snfrc" podStartSLOduration=5.643852293 podStartE2EDuration="1m41.880523056s" podCreationTimestamp="2025-12-09 16:59:24 +0000 UTC" firstStartedPulling="2025-12-09 16:59:28.7454946 +0000 UTC m=+165.133668420" lastFinishedPulling="2025-12-09 17:01:04.982165363 +0000 UTC m=+261.370339183" observedRunningTime="2025-12-09 17:01:05.846857788 +0000 UTC m=+262.235031618" watchObservedRunningTime="2025-12-09 17:01:05.880523056 +0000 UTC m=+262.268696876" Dec 09 17:01:05 crc kubenswrapper[4954]: I1209 17:01:05.949121 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j57d6" podStartSLOduration=5.707605324 podStartE2EDuration="1m41.949105783s" podCreationTimestamp="2025-12-09 16:59:24 +0000 UTC" firstStartedPulling="2025-12-09 16:59:28.734401914 +0000 UTC m=+165.122575734" lastFinishedPulling="2025-12-09 17:01:04.975902373 +0000 UTC m=+261.364076193" observedRunningTime="2025-12-09 17:01:05.880248177 +0000 UTC m=+262.268421997" watchObservedRunningTime="2025-12-09 17:01:05.949105783 +0000 UTC m=+262.337279603" Dec 09 17:01:05 crc kubenswrapper[4954]: I1209 17:01:05.949211 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c9xz9" podStartSLOduration=3.824166369 podStartE2EDuration="1m39.949208406s" podCreationTimestamp="2025-12-09 16:59:26 +0000 UTC" firstStartedPulling="2025-12-09 16:59:28.747350811 +0000 UTC m=+165.135524631" lastFinishedPulling="2025-12-09 17:01:04.872392848 +0000 UTC m=+261.260566668" observedRunningTime="2025-12-09 17:01:05.947698316 +0000 UTC m=+262.335872136" watchObservedRunningTime="2025-12-09 17:01:05.949208406 +0000 UTC m=+262.337382226" Dec 09 17:01:06 crc kubenswrapper[4954]: I1209 17:01:06.709587 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jjxbl" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerName="registry-server" probeResult="failure" output=< Dec 09 17:01:06 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 17:01:06 crc kubenswrapper[4954]: > Dec 09 17:01:06 crc kubenswrapper[4954]: I1209 17:01:06.711713 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-nlf5c" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerName="registry-server" probeResult="failure" output=< Dec 09 17:01:06 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 17:01:06 crc kubenswrapper[4954]: > Dec 09 17:01:06 crc kubenswrapper[4954]: I1209 17:01:06.943158 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7mwfq" event={"ID":"bcc59fe2-e410-4646-be11-560d0f45938b","Type":"ContainerStarted","Data":"e5fbba9ad18b1bb6fef18a9cf8bd59c22fa0955b79ee6e8bfb29ae7dfd46e8df"} Dec 09 17:01:06 crc kubenswrapper[4954]: I1209 17:01:06.965390 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7mwfq" podStartSLOduration=3.374888731 podStartE2EDuration="1m40.965374356s" podCreationTimestamp="2025-12-09 16:59:26 +0000 UTC" firstStartedPulling="2025-12-09 16:59:28.749898256 +0000 UTC m=+165.138072076" lastFinishedPulling="2025-12-09 17:01:06.340383881 +0000 UTC m=+262.728557701" observedRunningTime="2025-12-09 17:01:06.964196687 +0000 UTC m=+263.352370527" watchObservedRunningTime="2025-12-09 17:01:06.965374356 +0000 UTC m=+263.353548176" Dec 09 17:01:07 crc kubenswrapper[4954]: I1209 17:01:07.438024 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 17:01:07 crc kubenswrapper[4954]: I1209 17:01:07.438091 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 17:01:07 crc kubenswrapper[4954]: I1209 17:01:07.897159 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 17:01:07 crc kubenswrapper[4954]: I1209 17:01:07.897220 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 17:01:08 crc kubenswrapper[4954]: I1209 17:01:08.488487 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-c9xz9" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="registry-server" probeResult="failure" output=< Dec 09 17:01:08 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 17:01:08 crc kubenswrapper[4954]: > Dec 09 17:01:08 crc kubenswrapper[4954]: I1209 17:01:08.943335 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vfz9w" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerName="registry-server" probeResult="failure" output=< Dec 09 17:01:08 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 17:01:08 crc kubenswrapper[4954]: > Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.090717 4954 generic.go:334] "Generic (PLEG): container finished" podID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerID="55a050768b735b0c202117de2c9a3f0b7f1fec8a58d5438b40597da9a6adc931" exitCode=0 Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.090771 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpj95" event={"ID":"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56","Type":"ContainerDied","Data":"55a050768b735b0c202117de2c9a3f0b7f1fec8a58d5438b40597da9a6adc931"} Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.827045 4954 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.828198 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62" gracePeriod=15 Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.828194 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97" gracePeriod=15 Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.828198 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7" gracePeriod=15 Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.828177 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0" gracePeriod=15 Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.828287 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb" gracePeriod=15 Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.829482 4954 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 17:01:09 crc kubenswrapper[4954]: E1209 17:01:09.829817 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.829834 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 09 17:01:09 crc kubenswrapper[4954]: E1209 17:01:09.829848 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.829855 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 17:01:09 crc kubenswrapper[4954]: E1209 17:01:09.829869 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.829876 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 17:01:09 crc kubenswrapper[4954]: E1209 17:01:09.829886 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.829896 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 17:01:09 crc kubenswrapper[4954]: E1209 17:01:09.829906 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.829913 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 17:01:09 crc kubenswrapper[4954]: E1209 17:01:09.829923 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.829931 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 17:01:09 crc kubenswrapper[4954]: E1209 17:01:09.829946 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.829954 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.830095 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.830110 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.830122 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.830135 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.830145 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.830157 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.832885 4954 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.833895 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.839730 4954 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.880872 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.913890 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.913983 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.914038 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.914074 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.914109 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.914144 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.914178 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:09 crc kubenswrapper[4954]: I1209 17:01:09.914228 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.015891 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.015975 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016036 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016053 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016074 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016046 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016122 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016144 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016164 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016181 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016200 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016223 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016267 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016272 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016259 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.016238 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: I1209 17:01:10.176341 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:01:10 crc kubenswrapper[4954]: E1209 17:01:10.258519 4954 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f9ab277c91ce2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 17:01:10.257048802 +0000 UTC m=+266.645222622,LastTimestamp:2025-12-09 17:01:10.257048802 +0000 UTC m=+266.645222622,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.106905 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.108547 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.109288 4954 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97" exitCode=0 Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.109338 4954 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0" exitCode=0 Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.109351 4954 scope.go:117] "RemoveContainer" containerID="9683f809316393aecec95d1c96e68f282d0b942c0f60daaa8209b2579d4bbc2b" Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.109353 4954 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7" exitCode=0 Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.109431 4954 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb" exitCode=2 Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.111355 4954 generic.go:334] "Generic (PLEG): container finished" podID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" containerID="518372519c39eeb1f94b18331356f455b2c61dfa71a9fa822d1fa67c6cc7e937" exitCode=0 Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.111438 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ce53dd1a-0beb-4c91-9422-8473bd7a9412","Type":"ContainerDied","Data":"518372519c39eeb1f94b18331356f455b2c61dfa71a9fa822d1fa67c6cc7e937"} Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.115812 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.116290 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.116609 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34"} Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.116651 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"030b2a88a345965cd0fbfe0610c57a6f63b6fb03924ff6d926ccbeec99d4f79b"} Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.117279 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.117533 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.119208 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpj95" event={"ID":"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56","Type":"ContainerStarted","Data":"2b2d50b62b94bdd59bbb69d255d656e435b51c289d0ed62839023139383a3879"} Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.119678 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.120062 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:11 crc kubenswrapper[4954]: I1209 17:01:11.120468 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:11 crc kubenswrapper[4954]: E1209 17:01:11.615886 4954 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f9ab277c91ce2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 17:01:10.257048802 +0000 UTC m=+266.645222622,LastTimestamp:2025-12-09 17:01:10.257048802 +0000 UTC m=+266.645222622,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.126670 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.593255 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.595165 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.595576 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.595791 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.600506 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.601468 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.601902 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.602334 4954 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.602639 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.602892 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.748994 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.749172 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-var-lock\") pod \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.749229 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.749152 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.749276 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-var-lock" (OuterVolumeSpecName: "var-lock") pod "ce53dd1a-0beb-4c91-9422-8473bd7a9412" (UID: "ce53dd1a-0beb-4c91-9422-8473bd7a9412"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.749329 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.749337 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.749351 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.749396 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kube-api-access\") pod \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.749440 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kubelet-dir\") pod \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\" (UID: \"ce53dd1a-0beb-4c91-9422-8473bd7a9412\") " Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.749482 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ce53dd1a-0beb-4c91-9422-8473bd7a9412" (UID: "ce53dd1a-0beb-4c91-9422-8473bd7a9412"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.750002 4954 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.750018 4954 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-var-lock\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.750028 4954 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.750037 4954 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.750046 4954 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.756884 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ce53dd1a-0beb-4c91-9422-8473bd7a9412" (UID: "ce53dd1a-0beb-4c91-9422-8473bd7a9412"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:01:12 crc kubenswrapper[4954]: I1209 17:01:12.851514 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce53dd1a-0beb-4c91-9422-8473bd7a9412-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.140934 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.143560 4954 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62" exitCode=0 Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.143718 4954 scope.go:117] "RemoveContainer" containerID="63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.143862 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.148655 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ce53dd1a-0beb-4c91-9422-8473bd7a9412","Type":"ContainerDied","Data":"255c516d4832905ae7775e3be235f197ad3d6a0cc3957dd56c6810174fde4db7"} Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.148717 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="255c516d4832905ae7775e3be235f197ad3d6a0cc3957dd56c6810174fde4db7" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.148802 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.160559 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.160809 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.160996 4954 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.161171 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.173637 4954 scope.go:117] "RemoveContainer" containerID="163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.182989 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.183829 4954 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.184083 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.184278 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.204587 4954 scope.go:117] "RemoveContainer" containerID="813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.220593 4954 scope.go:117] "RemoveContainer" containerID="3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.236427 4954 scope.go:117] "RemoveContainer" containerID="65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.264041 4954 scope.go:117] "RemoveContainer" containerID="2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.291788 4954 scope.go:117] "RemoveContainer" containerID="63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.295798 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\": container with ID starting with 63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97 not found: ID does not exist" containerID="63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.295900 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97"} err="failed to get container status \"63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\": rpc error: code = NotFound desc = could not find container \"63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97\": container with ID starting with 63b4e2e49f9a224a7668ec5e90746ebf93b7ffa8ddbe937edf1d50501abbaf97 not found: ID does not exist" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.295958 4954 scope.go:117] "RemoveContainer" containerID="163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.296664 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\": container with ID starting with 163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0 not found: ID does not exist" containerID="163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.296717 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0"} err="failed to get container status \"163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\": rpc error: code = NotFound desc = could not find container \"163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0\": container with ID starting with 163f30b60a8ca4c46e2f7bb34162b6e6eb8cfeac44b71810c42318b5ad360fa0 not found: ID does not exist" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.296762 4954 scope.go:117] "RemoveContainer" containerID="813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.297168 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\": container with ID starting with 813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7 not found: ID does not exist" containerID="813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.297187 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7"} err="failed to get container status \"813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\": rpc error: code = NotFound desc = could not find container \"813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7\": container with ID starting with 813f4fd5a36fff835cc5766da908c2694b18f8d461c4fb97308ea4e3055c80a7 not found: ID does not exist" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.297203 4954 scope.go:117] "RemoveContainer" containerID="3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.297512 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\": container with ID starting with 3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb not found: ID does not exist" containerID="3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.297542 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb"} err="failed to get container status \"3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\": rpc error: code = NotFound desc = could not find container \"3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb\": container with ID starting with 3d948109174bd48a6d3aa9a64f6d0265ef82da2a2399b5246a1f44faa48f0feb not found: ID does not exist" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.297556 4954 scope.go:117] "RemoveContainer" containerID="65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.297842 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\": container with ID starting with 65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62 not found: ID does not exist" containerID="65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.297863 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62"} err="failed to get container status \"65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\": rpc error: code = NotFound desc = could not find container \"65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62\": container with ID starting with 65354d9d26bee6dfd8da987f21c601d753cc776e7a2cfbf4703c183a6ace0c62 not found: ID does not exist" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.297877 4954 scope.go:117] "RemoveContainer" containerID="2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.298086 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\": container with ID starting with 2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef not found: ID does not exist" containerID="2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.298109 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef"} err="failed to get container status \"2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\": rpc error: code = NotFound desc = could not find container \"2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef\": container with ID starting with 2fba3dcf0b8b469f3c8bfdff2f3bdde1b8ca5696425cddd6ae761c31e8d782ef not found: ID does not exist" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.728807 4954 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.729373 4954 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.729685 4954 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.729948 4954 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.730386 4954 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:13 crc kubenswrapper[4954]: I1209 17:01:13.730418 4954 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.730918 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Dec 09 17:01:13 crc kubenswrapper[4954]: E1209 17:01:13.931392 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Dec 09 17:01:14 crc kubenswrapper[4954]: I1209 17:01:14.123889 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:14 crc kubenswrapper[4954]: I1209 17:01:14.124363 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:14 crc kubenswrapper[4954]: I1209 17:01:14.124637 4954 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:14 crc kubenswrapper[4954]: I1209 17:01:14.124989 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:14 crc kubenswrapper[4954]: I1209 17:01:14.129066 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 09 17:01:14 crc kubenswrapper[4954]: E1209 17:01:14.332562 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Dec 09 17:01:15 crc kubenswrapper[4954]: E1209 17:01:15.132986 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.190189 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-snfrc" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.190248 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-snfrc" Dec 09 17:01:15 crc kubenswrapper[4954]: E1209 17:01:15.203332 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T17:01:10Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T17:01:10Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T17:01:10Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T17:01:10Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:a70fd7a02d90181eceb414b639a623f139139953c014ee2f3fcdf2fc98b7a02b\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:e5e992f87d6bb050c2c843099283c44e741d45fb6ea5530c8fda2ed27ed674fd\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1626040142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:27ad12b7f05c3823a316b4e5bcd7254f68d04a2d1e53b95f58fe6dafb7ee38d1\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:d0a5541748ed89ccd06733d73f1871982df0d544cfd91f5fc3e54e9807ade234\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1214998867},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1b1026c62413fa239fa4ff6541fe8bda656c1281867ad6ee2c848feccb13c97e\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:2b633ebdc901d19290af4dc2d09e2b59c504c0fc15a3fba410b0ce098e2d5753\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1141987142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: E1209 17:01:15.204086 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: E1209 17:01:15.204343 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: E1209 17:01:15.204652 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: E1209 17:01:15.204883 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: E1209 17:01:15.204906 4954 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.229160 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j57d6" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.229229 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j57d6" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.773023 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j57d6" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.773147 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.773520 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-snfrc" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.773531 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.773744 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.773973 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.774307 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.774507 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.774789 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.775262 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.776194 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.776308 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.776618 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.777012 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.777434 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.778194 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.778559 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.779208 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.779405 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.779559 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.779725 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.779993 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.816383 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.816791 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.817018 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.817328 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.817724 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.817780 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.818003 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.818306 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.818992 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.819754 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.820342 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.820686 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.821042 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.821359 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.821759 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:15 crc kubenswrapper[4954]: I1209 17:01:15.822179 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.213422 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j57d6" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.213968 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.214336 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.214703 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.214999 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.215268 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.215565 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.216070 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.220547 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-snfrc" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.220819 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.221054 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.221357 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.221577 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.221865 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.222077 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.222281 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: E1209 17:01:16.734806 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="3.2s" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.880976 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.881044 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.931632 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.932665 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.933176 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.933486 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.933855 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.934173 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.934501 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.934814 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:16 crc kubenswrapper[4954]: I1209 17:01:16.935158 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.223054 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.223731 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.224101 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.224837 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.225253 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.225559 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.225960 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.226279 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.226720 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.343847 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.344751 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.345267 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.348971 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.349556 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.349896 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.350220 4954 status_manager.go:851] "Failed to get status for pod" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" pod="openshift-marketplace/redhat-marketplace-c9xz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-c9xz9\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.350574 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.351253 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.351688 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.394274 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.395061 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.395440 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.395902 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.396164 4954 status_manager.go:851] "Failed to get status for pod" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" pod="openshift-marketplace/redhat-marketplace-c9xz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-c9xz9\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.396397 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.396820 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.397100 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.397430 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.397791 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.938235 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.939115 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.939625 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.939966 4954 status_manager.go:851] "Failed to get status for pod" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" pod="openshift-marketplace/redhat-operators-vfz9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vfz9w\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.940278 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.941138 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.941366 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.941537 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.941738 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.941898 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.942045 4954 status_manager.go:851] "Failed to get status for pod" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" pod="openshift-marketplace/redhat-marketplace-c9xz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-c9xz9\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.985726 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.986298 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.986989 4954 status_manager.go:851] "Failed to get status for pod" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" pod="openshift-marketplace/redhat-marketplace-c9xz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-c9xz9\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.987767 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.988058 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.988345 4954 status_manager.go:851] "Failed to get status for pod" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" pod="openshift-marketplace/redhat-operators-vfz9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vfz9w\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.988582 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.989050 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.989737 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.990125 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:17 crc kubenswrapper[4954]: I1209 17:01:17.990562 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.355968 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.356051 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.416960 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.417768 4954 status_manager.go:851] "Failed to get status for pod" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" pod="openshift-marketplace/redhat-operators-vfz9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vfz9w\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.418500 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.419145 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.419444 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.419807 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.420156 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.420531 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.421281 4954 status_manager.go:851] "Failed to get status for pod" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" pod="openshift-marketplace/redhat-marketplace-c9xz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-c9xz9\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.421795 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:18 crc kubenswrapper[4954]: I1209 17:01:18.422088 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.229389 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.230107 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.230462 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.231087 4954 status_manager.go:851] "Failed to get status for pod" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" pod="openshift-marketplace/redhat-marketplace-c9xz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-c9xz9\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.231560 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.231916 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.232267 4954 status_manager.go:851] "Failed to get status for pod" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" pod="openshift-marketplace/redhat-operators-vfz9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vfz9w\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.232676 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.233187 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.233541 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: I1209 17:01:19.233930 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:19 crc kubenswrapper[4954]: E1209 17:01:19.936310 4954 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="6.4s" Dec 09 17:01:21 crc kubenswrapper[4954]: E1209 17:01:21.616928 4954 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f9ab277c91ce2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 17:01:10.257048802 +0000 UTC m=+266.645222622,LastTimestamp:2025-12-09 17:01:10.257048802 +0000 UTC m=+266.645222622,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.119950 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.121250 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.121734 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.122251 4954 status_manager.go:851] "Failed to get status for pod" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" pod="openshift-marketplace/redhat-operators-vfz9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vfz9w\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.122572 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.122843 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.123328 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.124217 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.124635 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.124973 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.125276 4954 status_manager.go:851] "Failed to get status for pod" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" pod="openshift-marketplace/redhat-marketplace-c9xz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-c9xz9\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.147576 4954 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.147644 4954 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:23 crc kubenswrapper[4954]: E1209 17:01:23.148612 4954 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.149201 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:23 crc kubenswrapper[4954]: W1209 17:01:23.180899 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-f421b1799e656d1b3e20e8294f808cf5c5a54a10bd4d3f2f989c5ed023e4aa59 WatchSource:0}: Error finding container f421b1799e656d1b3e20e8294f808cf5c5a54a10bd4d3f2f989c5ed023e4aa59: Status 404 returned error can't find the container with id f421b1799e656d1b3e20e8294f808cf5c5a54a10bd4d3f2f989c5ed023e4aa59 Dec 09 17:01:23 crc kubenswrapper[4954]: I1209 17:01:23.212221 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f421b1799e656d1b3e20e8294f808cf5c5a54a10bd4d3f2f989c5ed023e4aa59"} Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.126920 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.127667 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.127921 4954 status_manager.go:851] "Failed to get status for pod" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" pod="openshift-marketplace/redhat-operators-vfz9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vfz9w\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.128408 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.129194 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.129540 4954 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.130032 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.130437 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.130757 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.131071 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:24 crc kubenswrapper[4954]: I1209 17:01:24.131437 4954 status_manager.go:851] "Failed to get status for pod" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" pod="openshift-marketplace/redhat-marketplace-c9xz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-c9xz9\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.230481 4954 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="e0c33b2ac881a7ed25caae882ca0620a9f2c5ce7f84971ddbdc5745fdaddf62d" exitCode=0 Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.231112 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"e0c33b2ac881a7ed25caae882ca0620a9f2c5ce7f84971ddbdc5745fdaddf62d"} Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.231461 4954 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.233110 4954 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.232091 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: E1209 17:01:25.233633 4954 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.233750 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.233924 4954 status_manager.go:851] "Failed to get status for pod" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" pod="openshift-marketplace/redhat-operators-vfz9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vfz9w\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.234114 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.234286 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.234443 4954 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.234645 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.234810 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.235133 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.235661 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.235884 4954 status_manager.go:851] "Failed to get status for pod" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" pod="openshift-marketplace/redhat-marketplace-c9xz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-c9xz9\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.238805 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.238865 4954 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f" exitCode=1 Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.238908 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f"} Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.239700 4954 scope.go:117] "RemoveContainer" containerID="a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.239906 4954 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.240138 4954 status_manager.go:851] "Failed to get status for pod" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" pod="openshift-marketplace/certified-operators-jjxbl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-jjxbl\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.240413 4954 status_manager.go:851] "Failed to get status for pod" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" pod="openshift-marketplace/community-operators-j57d6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-j57d6\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.240685 4954 status_manager.go:851] "Failed to get status for pod" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" pod="openshift-marketplace/redhat-marketplace-c9xz9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-c9xz9\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.240904 4954 status_manager.go:851] "Failed to get status for pod" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" pod="openshift-marketplace/redhat-marketplace-7mwfq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-7mwfq\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.241105 4954 status_manager.go:851] "Failed to get status for pod" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" pod="openshift-marketplace/redhat-operators-wpj95" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wpj95\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.241379 4954 status_manager.go:851] "Failed to get status for pod" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" pod="openshift-marketplace/redhat-operators-vfz9w" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-vfz9w\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.241806 4954 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.242185 4954 status_manager.go:851] "Failed to get status for pod" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" pod="openshift-marketplace/certified-operators-nlf5c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-nlf5c\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.242458 4954 status_manager.go:851] "Failed to get status for pod" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" pod="openshift-marketplace/community-operators-snfrc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-snfrc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.242724 4954 status_manager.go:851] "Failed to get status for pod" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: I1209 17:01:25.242976 4954 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: E1209 17:01:25.497871 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T17:01:25Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T17:01:25Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T17:01:25Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T17:01:25Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:a70fd7a02d90181eceb414b639a623f139139953c014ee2f3fcdf2fc98b7a02b\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:e5e992f87d6bb050c2c843099283c44e741d45fb6ea5530c8fda2ed27ed674fd\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1626040142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:27ad12b7f05c3823a316b4e5bcd7254f68d04a2d1e53b95f58fe6dafb7ee38d1\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:d0a5541748ed89ccd06733d73f1871982df0d544cfd91f5fc3e54e9807ade234\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1214998867},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1b1026c62413fa239fa4ff6541fe8bda656c1281867ad6ee2c848feccb13c97e\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:2b633ebdc901d19290af4dc2d09e2b59c504c0fc15a3fba410b0ce098e2d5753\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1141987142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: E1209 17:01:25.498475 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: E1209 17:01:25.499011 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: E1209 17:01:25.499280 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: E1209 17:01:25.499511 4954 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Dec 09 17:01:25 crc kubenswrapper[4954]: E1209 17:01:25.499537 4954 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 17:01:26 crc kubenswrapper[4954]: I1209 17:01:26.255431 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a1464e8cb97675b2a12f134581fd405c3c3b5e016ce020de169a17df84df5717"} Dec 09 17:01:26 crc kubenswrapper[4954]: I1209 17:01:26.255727 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"423d96dd3278aecd147695db90ea729f4f10cbf954349a847d6049c75c1cea57"} Dec 09 17:01:26 crc kubenswrapper[4954]: I1209 17:01:26.255737 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"edd5f168986e2e72c914a2ef0d296938b103674d6e6e7ea5fc32ab7159605a70"} Dec 09 17:01:26 crc kubenswrapper[4954]: I1209 17:01:26.255749 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"91129564a8c1a94f9afe9a07f52d5c4391666edc81daad0ce351135ea6fe8500"} Dec 09 17:01:26 crc kubenswrapper[4954]: I1209 17:01:26.269306 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 09 17:01:26 crc kubenswrapper[4954]: I1209 17:01:26.269376 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0d96526802b83b303759cac50ad985dea04fd0a6d5b3f45b83d3be452a1d1a43"} Dec 09 17:01:27 crc kubenswrapper[4954]: I1209 17:01:27.277909 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4da9adde677b10f3c68296010702722ca78ecaf681c1fc6fa19914d864c8a782"} Dec 09 17:01:27 crc kubenswrapper[4954]: I1209 17:01:27.278370 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:27 crc kubenswrapper[4954]: I1209 17:01:27.278194 4954 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:27 crc kubenswrapper[4954]: I1209 17:01:27.278391 4954 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:27 crc kubenswrapper[4954]: I1209 17:01:27.553431 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 17:01:27 crc kubenswrapper[4954]: I1209 17:01:27.553670 4954 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 09 17:01:27 crc kubenswrapper[4954]: I1209 17:01:27.553722 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 09 17:01:28 crc kubenswrapper[4954]: I1209 17:01:28.150366 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:28 crc kubenswrapper[4954]: I1209 17:01:28.150783 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:28 crc kubenswrapper[4954]: I1209 17:01:28.156060 4954 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]log ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]etcd ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/generic-apiserver-start-informers ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/priority-and-fairness-filter ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/start-apiextensions-informers ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/start-apiextensions-controllers ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/crd-informer-synced ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/start-system-namespaces-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 09 17:01:28 crc kubenswrapper[4954]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/bootstrap-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/start-kube-aggregator-informers ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/apiservice-registration-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/apiservice-discovery-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]autoregister-completion ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/apiservice-openapi-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 09 17:01:28 crc kubenswrapper[4954]: livez check failed Dec 09 17:01:28 crc kubenswrapper[4954]: I1209 17:01:28.156126 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 17:01:29 crc kubenswrapper[4954]: I1209 17:01:29.218423 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 17:01:32 crc kubenswrapper[4954]: I1209 17:01:32.288455 4954 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:33 crc kubenswrapper[4954]: I1209 17:01:33.156332 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:33 crc kubenswrapper[4954]: I1209 17:01:33.158850 4954 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="46bf5c1a-638f-4d07-b75b-fe4979dd1dca" Dec 09 17:01:33 crc kubenswrapper[4954]: I1209 17:01:33.311422 4954 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:33 crc kubenswrapper[4954]: I1209 17:01:33.311466 4954 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:33 crc kubenswrapper[4954]: I1209 17:01:33.326557 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:01:34 crc kubenswrapper[4954]: I1209 17:01:34.142230 4954 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="46bf5c1a-638f-4d07-b75b-fe4979dd1dca" Dec 09 17:01:34 crc kubenswrapper[4954]: I1209 17:01:34.322738 4954 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:34 crc kubenswrapper[4954]: I1209 17:01:34.322774 4954 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:34 crc kubenswrapper[4954]: I1209 17:01:34.326328 4954 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="46bf5c1a-638f-4d07-b75b-fe4979dd1dca" Dec 09 17:01:35 crc kubenswrapper[4954]: I1209 17:01:35.327067 4954 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:35 crc kubenswrapper[4954]: I1209 17:01:35.327413 4954 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="06b58b47-128b-4499-adcb-16e004b35592" Dec 09 17:01:35 crc kubenswrapper[4954]: I1209 17:01:35.330813 4954 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="46bf5c1a-638f-4d07-b75b-fe4979dd1dca" Dec 09 17:01:37 crc kubenswrapper[4954]: I1209 17:01:37.554158 4954 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 09 17:01:37 crc kubenswrapper[4954]: I1209 17:01:37.554227 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 09 17:01:42 crc kubenswrapper[4954]: I1209 17:01:42.084464 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 09 17:01:42 crc kubenswrapper[4954]: I1209 17:01:42.325932 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 09 17:01:42 crc kubenswrapper[4954]: I1209 17:01:42.432166 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 09 17:01:42 crc kubenswrapper[4954]: I1209 17:01:42.548989 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 09 17:01:42 crc kubenswrapper[4954]: I1209 17:01:42.631217 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 09 17:01:42 crc kubenswrapper[4954]: I1209 17:01:42.715763 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 09 17:01:42 crc kubenswrapper[4954]: I1209 17:01:42.975934 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 09 17:01:43 crc kubenswrapper[4954]: I1209 17:01:43.339636 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 09 17:01:43 crc kubenswrapper[4954]: I1209 17:01:43.787536 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 09 17:01:44 crc kubenswrapper[4954]: I1209 17:01:44.114947 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 09 17:01:44 crc kubenswrapper[4954]: I1209 17:01:44.206914 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 09 17:01:44 crc kubenswrapper[4954]: I1209 17:01:44.378699 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 17:01:44 crc kubenswrapper[4954]: I1209 17:01:44.439943 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 09 17:01:44 crc kubenswrapper[4954]: I1209 17:01:44.575253 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 09 17:01:44 crc kubenswrapper[4954]: I1209 17:01:44.749719 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 09 17:01:44 crc kubenswrapper[4954]: I1209 17:01:44.868093 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 09 17:01:44 crc kubenswrapper[4954]: I1209 17:01:44.997661 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.005049 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.015640 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.062454 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.241512 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.278886 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.355862 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.609144 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.609655 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.656540 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.720956 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.760234 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.829286 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 09 17:01:45 crc kubenswrapper[4954]: I1209 17:01:45.842143 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.037577 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.122342 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.239128 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.306583 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.346129 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.367703 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.448752 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.454894 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.481383 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.525457 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.583851 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.601713 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.633757 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.760820 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.776791 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.782765 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.833255 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.834073 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.886880 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.906211 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 09 17:01:46 crc kubenswrapper[4954]: I1209 17:01:46.987490 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.030436 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.143870 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.208049 4954 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.216188 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.284390 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.324436 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.395060 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.437536 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.453843 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.499367 4954 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.508806 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.554299 4954 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.554691 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.554880 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.555691 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"0d96526802b83b303759cac50ad985dea04fd0a6d5b3f45b83d3be452a1d1a43"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.555929 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://0d96526802b83b303759cac50ad985dea04fd0a6d5b3f45b83d3be452a1d1a43" gracePeriod=30 Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.592166 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.619287 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.625481 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.745486 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.745518 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.838086 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.899354 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 09 17:01:47 crc kubenswrapper[4954]: I1209 17:01:47.961522 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.191340 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.245682 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.284158 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.293754 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.353304 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.361372 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.425350 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.430032 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.488046 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.649343 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.653284 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.675332 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.718312 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.718755 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.733312 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.749045 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.749255 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.752268 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.821658 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.943007 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.962184 4954 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 09 17:01:48 crc kubenswrapper[4954]: I1209 17:01:48.988818 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.040185 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.075269 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.174407 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.223370 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.239193 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.316934 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.324092 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.352926 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.360663 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.382444 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.455170 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.540277 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.700520 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.701163 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.848908 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.928980 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 09 17:01:49 crc kubenswrapper[4954]: I1209 17:01:49.946500 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.129903 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.130675 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.203298 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.345617 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.410688 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.453529 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.477720 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.481440 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.510673 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.521884 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.573078 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.648360 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.662722 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.684345 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.737491 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.749499 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.840570 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.841325 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 09 17:01:50 crc kubenswrapper[4954]: I1209 17:01:50.907339 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.104633 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.121682 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.134224 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.172513 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.237696 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.292678 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.357190 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.395233 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.460814 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.474414 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.527195 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.536811 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.538964 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.597543 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.693785 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.833197 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.940714 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.966232 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 09 17:01:51 crc kubenswrapper[4954]: I1209 17:01:51.987542 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.102801 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.140948 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.145961 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.154995 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.171776 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.241456 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.245450 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.249909 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.274303 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.305061 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.336905 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.372099 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.407903 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.486950 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.526425 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.527678 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.548185 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.604866 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.638509 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.697548 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.716400 4954 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.716959 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.746680 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.806268 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.873766 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.891660 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 09 17:01:52 crc kubenswrapper[4954]: I1209 17:01:52.918913 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.023869 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.035291 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.142497 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.167060 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.207006 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.226049 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.255234 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.268774 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.304005 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.444308 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.473630 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.615380 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.627983 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.680430 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.725464 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.774730 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.873673 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 09 17:01:53 crc kubenswrapper[4954]: I1209 17:01:53.894099 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.087529 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.099065 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.126511 4954 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.144324 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.302632 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.376609 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.383247 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.487861 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.553836 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.675831 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.676047 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.703727 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.757342 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.780090 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.849922 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.863939 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 09 17:01:54 crc kubenswrapper[4954]: I1209 17:01:54.985695 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.007717 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.077198 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.095051 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.250700 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.324640 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.432747 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.436852 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.445159 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.492936 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.645776 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.696827 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.714092 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.739302 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.787716 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.896202 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.922173 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 09 17:01:55 crc kubenswrapper[4954]: I1209 17:01:55.967253 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 09 17:01:56 crc kubenswrapper[4954]: I1209 17:01:56.004049 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 09 17:01:56 crc kubenswrapper[4954]: I1209 17:01:56.005867 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 09 17:01:56 crc kubenswrapper[4954]: I1209 17:01:56.009704 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 09 17:01:56 crc kubenswrapper[4954]: I1209 17:01:56.100621 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 09 17:01:56 crc kubenswrapper[4954]: I1209 17:01:56.213258 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 17:01:56 crc kubenswrapper[4954]: I1209 17:01:56.266650 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 17:01:56 crc kubenswrapper[4954]: I1209 17:01:56.326107 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 09 17:01:56 crc kubenswrapper[4954]: I1209 17:01:56.477361 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 09 17:01:56 crc kubenswrapper[4954]: I1209 17:01:56.536186 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 09 17:01:56 crc kubenswrapper[4954]: I1209 17:01:56.931323 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.063999 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.064338 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.119778 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.169414 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.234873 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.280197 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.305387 4954 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.307075 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wpj95" podStartSLOduration=51.101581306 podStartE2EDuration="2m30.307050965s" podCreationTimestamp="2025-12-09 16:59:27 +0000 UTC" firstStartedPulling="2025-12-09 16:59:31.078861297 +0000 UTC m=+167.467035127" lastFinishedPulling="2025-12-09 17:01:10.284330966 +0000 UTC m=+266.672504786" observedRunningTime="2025-12-09 17:01:32.069576703 +0000 UTC m=+288.457750523" watchObservedRunningTime="2025-12-09 17:01:57.307050965 +0000 UTC m=+313.695224785" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.308564 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=48.308557245 podStartE2EDuration="48.308557245s" podCreationTimestamp="2025-12-09 17:01:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:01:32.216201045 +0000 UTC m=+288.604374865" watchObservedRunningTime="2025-12-09 17:01:57.308557245 +0000 UTC m=+313.696731065" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.310722 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.310777 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.310802 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j57d6","openshift-marketplace/marketplace-operator-79b997595-8jkvw","openshift-marketplace/certified-operators-jjxbl","openshift-marketplace/redhat-operators-wpj95","openshift-marketplace/redhat-operators-vfz9w","openshift-marketplace/redhat-marketplace-c9xz9","openshift-marketplace/community-operators-snfrc","openshift-marketplace/redhat-marketplace-7mwfq","openshift-marketplace/certified-operators-nlf5c"] Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.311087 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nlf5c" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerName="registry-server" containerID="cri-o://76b76e8114b08f3a9678e802e06fa776197be9e1d949be861a4bbe1b9688c346" gracePeriod=30 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.311242 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7mwfq" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" containerName="registry-server" containerID="cri-o://e5fbba9ad18b1bb6fef18a9cf8bd59c22fa0955b79ee6e8bfb29ae7dfd46e8df" gracePeriod=30 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.311522 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j57d6" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerName="registry-server" containerID="cri-o://474a91cd7207e014c2d4d28ffbb23669f09afd2ee82fd6fcde0d2bfafbe7845d" gracePeriod=30 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.311851 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-snfrc" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerName="registry-server" containerID="cri-o://62ba3752232989ae967c4086a4ac4d99e2ef36ed7c7dcd22835f3ba01353318b" gracePeriod=30 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.312095 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" podUID="454e49f6-b054-4901-bd18-461d7e9cd89e" containerName="marketplace-operator" containerID="cri-o://5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13" gracePeriod=30 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.313193 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wpj95" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerName="registry-server" containerID="cri-o://2b2d50b62b94bdd59bbb69d255d656e435b51c289d0ed62839023139383a3879" gracePeriod=30 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.313367 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vfz9w" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerName="registry-server" containerID="cri-o://d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8" gracePeriod=30 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.313530 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c9xz9" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="registry-server" containerID="cri-o://975ac2f9c9073ce03cdb134f155cb901b644766788af76e2c3348b40c59adfed" gracePeriod=30 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.313718 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jjxbl" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerName="registry-server" containerID="cri-o://b67d954d432fdaefd19cddd78c8dc54206aa1618d67cb9535fc075be133e30b2" gracePeriod=30 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.360029 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-c9xz9" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="registry-server" probeResult="failure" output="" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.365938 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=25.365919547 podStartE2EDuration="25.365919547s" podCreationTimestamp="2025-12-09 17:01:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:01:57.364979925 +0000 UTC m=+313.753153755" watchObservedRunningTime="2025-12-09 17:01:57.365919547 +0000 UTC m=+313.754093367" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.366656 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-c9xz9" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="registry-server" probeResult="failure" output="" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.457771 4954 generic.go:334] "Generic (PLEG): container finished" podID="bcc59fe2-e410-4646-be11-560d0f45938b" containerID="e5fbba9ad18b1bb6fef18a9cf8bd59c22fa0955b79ee6e8bfb29ae7dfd46e8df" exitCode=0 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.457824 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7mwfq" event={"ID":"bcc59fe2-e410-4646-be11-560d0f45938b","Type":"ContainerDied","Data":"e5fbba9ad18b1bb6fef18a9cf8bd59c22fa0955b79ee6e8bfb29ae7dfd46e8df"} Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.461531 4954 generic.go:334] "Generic (PLEG): container finished" podID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerID="d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8" exitCode=0 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.461938 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vfz9w" event={"ID":"d0dc2999-1020-4b4b-96ae-87ba32e34ad8","Type":"ContainerDied","Data":"d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8"} Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.469076 4954 generic.go:334] "Generic (PLEG): container finished" podID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerID="474a91cd7207e014c2d4d28ffbb23669f09afd2ee82fd6fcde0d2bfafbe7845d" exitCode=0 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.469300 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57d6" event={"ID":"f28a7275-4712-4dc9-89a3-16a77bba72d1","Type":"ContainerDied","Data":"474a91cd7207e014c2d4d28ffbb23669f09afd2ee82fd6fcde0d2bfafbe7845d"} Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.472516 4954 generic.go:334] "Generic (PLEG): container finished" podID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerID="b67d954d432fdaefd19cddd78c8dc54206aa1618d67cb9535fc075be133e30b2" exitCode=0 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.472779 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjxbl" event={"ID":"c9ebbd26-bcdc-4db3-b345-a5715ac59a82","Type":"ContainerDied","Data":"b67d954d432fdaefd19cddd78c8dc54206aa1618d67cb9535fc075be133e30b2"} Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.473711 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.479840 4954 generic.go:334] "Generic (PLEG): container finished" podID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerID="2b2d50b62b94bdd59bbb69d255d656e435b51c289d0ed62839023139383a3879" exitCode=0 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.480075 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpj95" event={"ID":"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56","Type":"ContainerDied","Data":"2b2d50b62b94bdd59bbb69d255d656e435b51c289d0ed62839023139383a3879"} Dec 09 17:01:57 crc kubenswrapper[4954]: E1209 17:01:57.482654 4954 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9ebbd26_bcdc_4db3_b345_a5715ac59a82.slice/crio-b67d954d432fdaefd19cddd78c8dc54206aa1618d67cb9535fc075be133e30b2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5e4dfeb_181d_43c7_89bb_3c828fb5ce56.slice/crio-2b2d50b62b94bdd59bbb69d255d656e435b51c289d0ed62839023139383a3879.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod454e49f6_b054_4901_bd18_461d7e9cd89e.slice/crio-conmon-5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5e4dfeb_181d_43c7_89bb_3c828fb5ce56.slice/crio-conmon-2b2d50b62b94bdd59bbb69d255d656e435b51c289d0ed62839023139383a3879.scope\": RecentStats: unable to find data in memory cache]" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.504390 4954 generic.go:334] "Generic (PLEG): container finished" podID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerID="76b76e8114b08f3a9678e802e06fa776197be9e1d949be861a4bbe1b9688c346" exitCode=0 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.504477 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nlf5c" event={"ID":"271559e2-f374-4aeb-a3f7-6b039afc7e8f","Type":"ContainerDied","Data":"76b76e8114b08f3a9678e802e06fa776197be9e1d949be861a4bbe1b9688c346"} Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.523422 4954 generic.go:334] "Generic (PLEG): container finished" podID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerID="62ba3752232989ae967c4086a4ac4d99e2ef36ed7c7dcd22835f3ba01353318b" exitCode=0 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.523698 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-snfrc" event={"ID":"1df4d5f7-03d7-4b91-b949-b555dae7ce2b","Type":"ContainerDied","Data":"62ba3752232989ae967c4086a4ac4d99e2ef36ed7c7dcd22835f3ba01353318b"} Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.530887 4954 generic.go:334] "Generic (PLEG): container finished" podID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerID="975ac2f9c9073ce03cdb134f155cb901b644766788af76e2c3348b40c59adfed" exitCode=0 Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.530935 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9xz9" event={"ID":"f344130d-4bc2-4575-bb60-815ea6ae6a84","Type":"ContainerDied","Data":"975ac2f9c9073ce03cdb134f155cb901b644766788af76e2c3348b40c59adfed"} Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.666961 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.679221 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.750064 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 09 17:01:57 crc kubenswrapper[4954]: I1209 17:01:57.891422 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 09 17:01:57 crc kubenswrapper[4954]: E1209 17:01:57.898673 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8 is running failed: container process not found" containerID="d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 17:01:57 crc kubenswrapper[4954]: E1209 17:01:57.899370 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8 is running failed: container process not found" containerID="d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 17:01:57 crc kubenswrapper[4954]: E1209 17:01:57.899764 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8 is running failed: container process not found" containerID="d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 17:01:57 crc kubenswrapper[4954]: E1209 17:01:57.899792 4954 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-vfz9w" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerName="registry-server" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.017257 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.022432 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j57d6" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.026721 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-snfrc" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.032115 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.034815 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-catalog-content\") pod \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.034885 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-catalog-content\") pod \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.034938 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-utilities\") pod \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.034970 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-utilities\") pod \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.034997 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-catalog-content\") pod \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.035028 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-utilities\") pod \"f28a7275-4712-4dc9-89a3-16a77bba72d1\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.035059 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pzhr\" (UniqueName: \"kubernetes.io/projected/271559e2-f374-4aeb-a3f7-6b039afc7e8f-kube-api-access-9pzhr\") pod \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.036129 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-utilities" (OuterVolumeSpecName: "utilities") pod "1df4d5f7-03d7-4b91-b949-b555dae7ce2b" (UID: "1df4d5f7-03d7-4b91-b949-b555dae7ce2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.036465 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-utilities" (OuterVolumeSpecName: "utilities") pod "f28a7275-4712-4dc9-89a3-16a77bba72d1" (UID: "f28a7275-4712-4dc9-89a3-16a77bba72d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.036541 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-utilities" (OuterVolumeSpecName: "utilities") pod "c9ebbd26-bcdc-4db3-b345-a5715ac59a82" (UID: "c9ebbd26-bcdc-4db3-b345-a5715ac59a82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.038155 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmdfn\" (UniqueName: \"kubernetes.io/projected/f28a7275-4712-4dc9-89a3-16a77bba72d1-kube-api-access-nmdfn\") pod \"f28a7275-4712-4dc9-89a3-16a77bba72d1\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.038211 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-catalog-content\") pod \"f28a7275-4712-4dc9-89a3-16a77bba72d1\" (UID: \"f28a7275-4712-4dc9-89a3-16a77bba72d1\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.038251 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zdx7\" (UniqueName: \"kubernetes.io/projected/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-kube-api-access-2zdx7\") pod \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\" (UID: \"1df4d5f7-03d7-4b91-b949-b555dae7ce2b\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.038278 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlcll\" (UniqueName: \"kubernetes.io/projected/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-kube-api-access-qlcll\") pod \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\" (UID: \"c9ebbd26-bcdc-4db3-b345-a5715ac59a82\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.038315 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-utilities\") pod \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\" (UID: \"271559e2-f374-4aeb-a3f7-6b039afc7e8f\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.038777 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.038808 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.038822 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.039209 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.040380 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-utilities" (OuterVolumeSpecName: "utilities") pod "271559e2-f374-4aeb-a3f7-6b039afc7e8f" (UID: "271559e2-f374-4aeb-a3f7-6b039afc7e8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.042797 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/271559e2-f374-4aeb-a3f7-6b039afc7e8f-kube-api-access-9pzhr" (OuterVolumeSpecName: "kube-api-access-9pzhr") pod "271559e2-f374-4aeb-a3f7-6b039afc7e8f" (UID: "271559e2-f374-4aeb-a3f7-6b039afc7e8f"). InnerVolumeSpecName "kube-api-access-9pzhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.043632 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-kube-api-access-2zdx7" (OuterVolumeSpecName: "kube-api-access-2zdx7") pod "1df4d5f7-03d7-4b91-b949-b555dae7ce2b" (UID: "1df4d5f7-03d7-4b91-b949-b555dae7ce2b"). InnerVolumeSpecName "kube-api-access-2zdx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.045254 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f28a7275-4712-4dc9-89a3-16a77bba72d1-kube-api-access-nmdfn" (OuterVolumeSpecName: "kube-api-access-nmdfn") pod "f28a7275-4712-4dc9-89a3-16a77bba72d1" (UID: "f28a7275-4712-4dc9-89a3-16a77bba72d1"). InnerVolumeSpecName "kube-api-access-nmdfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.044967 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-kube-api-access-qlcll" (OuterVolumeSpecName: "kube-api-access-qlcll") pod "c9ebbd26-bcdc-4db3-b345-a5715ac59a82" (UID: "c9ebbd26-bcdc-4db3-b345-a5715ac59a82"). InnerVolumeSpecName "kube-api-access-qlcll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.050375 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.072949 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.098712 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.100611 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.111181 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9ebbd26-bcdc-4db3-b345-a5715ac59a82" (UID: "c9ebbd26-bcdc-4db3-b345-a5715ac59a82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.124826 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f28a7275-4712-4dc9-89a3-16a77bba72d1" (UID: "f28a7275-4712-4dc9-89a3-16a77bba72d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.125342 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "271559e2-f374-4aeb-a3f7-6b039afc7e8f" (UID: "271559e2-f374-4aeb-a3f7-6b039afc7e8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.125587 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1df4d5f7-03d7-4b91-b949-b555dae7ce2b" (UID: "1df4d5f7-03d7-4b91-b949-b555dae7ce2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.139955 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlcll\" (UniqueName: \"kubernetes.io/projected/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-kube-api-access-qlcll\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.140012 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.140029 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/271559e2-f374-4aeb-a3f7-6b039afc7e8f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.140041 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ebbd26-bcdc-4db3-b345-a5715ac59a82-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.140053 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.140066 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pzhr\" (UniqueName: \"kubernetes.io/projected/271559e2-f374-4aeb-a3f7-6b039afc7e8f-kube-api-access-9pzhr\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.140079 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmdfn\" (UniqueName: \"kubernetes.io/projected/f28a7275-4712-4dc9-89a3-16a77bba72d1-kube-api-access-nmdfn\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.140121 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f28a7275-4712-4dc9-89a3-16a77bba72d1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.140134 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zdx7\" (UniqueName: \"kubernetes.io/projected/1df4d5f7-03d7-4b91-b949-b555dae7ce2b-kube-api-access-2zdx7\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.240903 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-operator-metrics\") pod \"454e49f6-b054-4901-bd18-461d7e9cd89e\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.240963 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-utilities\") pod \"f344130d-4bc2-4575-bb60-815ea6ae6a84\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.240990 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnqqf\" (UniqueName: \"kubernetes.io/projected/454e49f6-b054-4901-bd18-461d7e9cd89e-kube-api-access-rnqqf\") pod \"454e49f6-b054-4901-bd18-461d7e9cd89e\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241016 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-trusted-ca\") pod \"454e49f6-b054-4901-bd18-461d7e9cd89e\" (UID: \"454e49f6-b054-4901-bd18-461d7e9cd89e\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241038 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjprj\" (UniqueName: \"kubernetes.io/projected/f344130d-4bc2-4575-bb60-815ea6ae6a84-kube-api-access-sjprj\") pod \"f344130d-4bc2-4575-bb60-815ea6ae6a84\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241062 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjj6v\" (UniqueName: \"kubernetes.io/projected/bcc59fe2-e410-4646-be11-560d0f45938b-kube-api-access-sjj6v\") pod \"bcc59fe2-e410-4646-be11-560d0f45938b\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241108 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-utilities\") pod \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241131 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-catalog-content\") pod \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241157 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-utilities\") pod \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241173 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-catalog-content\") pod \"f344130d-4bc2-4575-bb60-815ea6ae6a84\" (UID: \"f344130d-4bc2-4575-bb60-815ea6ae6a84\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241209 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-catalog-content\") pod \"bcc59fe2-e410-4646-be11-560d0f45938b\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241245 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-catalog-content\") pod \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241274 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-utilities\") pod \"bcc59fe2-e410-4646-be11-560d0f45938b\" (UID: \"bcc59fe2-e410-4646-be11-560d0f45938b\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241307 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mnvn\" (UniqueName: \"kubernetes.io/projected/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-kube-api-access-7mnvn\") pod \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\" (UID: \"d0dc2999-1020-4b4b-96ae-87ba32e34ad8\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.241323 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vvfc\" (UniqueName: \"kubernetes.io/projected/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-kube-api-access-9vvfc\") pod \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\" (UID: \"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56\") " Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.242082 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "454e49f6-b054-4901-bd18-461d7e9cd89e" (UID: "454e49f6-b054-4901-bd18-461d7e9cd89e"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.242486 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-utilities" (OuterVolumeSpecName: "utilities") pod "bcc59fe2-e410-4646-be11-560d0f45938b" (UID: "bcc59fe2-e410-4646-be11-560d0f45938b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.242996 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-utilities" (OuterVolumeSpecName: "utilities") pod "f344130d-4bc2-4575-bb60-815ea6ae6a84" (UID: "f344130d-4bc2-4575-bb60-815ea6ae6a84"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.243553 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-utilities" (OuterVolumeSpecName: "utilities") pod "a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" (UID: "a5e4dfeb-181d-43c7-89bb-3c828fb5ce56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.243778 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-utilities" (OuterVolumeSpecName: "utilities") pod "d0dc2999-1020-4b4b-96ae-87ba32e34ad8" (UID: "d0dc2999-1020-4b4b-96ae-87ba32e34ad8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.244558 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "454e49f6-b054-4901-bd18-461d7e9cd89e" (UID: "454e49f6-b054-4901-bd18-461d7e9cd89e"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.244944 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcc59fe2-e410-4646-be11-560d0f45938b-kube-api-access-sjj6v" (OuterVolumeSpecName: "kube-api-access-sjj6v") pod "bcc59fe2-e410-4646-be11-560d0f45938b" (UID: "bcc59fe2-e410-4646-be11-560d0f45938b"). InnerVolumeSpecName "kube-api-access-sjj6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.245409 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-kube-api-access-7mnvn" (OuterVolumeSpecName: "kube-api-access-7mnvn") pod "d0dc2999-1020-4b4b-96ae-87ba32e34ad8" (UID: "d0dc2999-1020-4b4b-96ae-87ba32e34ad8"). InnerVolumeSpecName "kube-api-access-7mnvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.246192 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f344130d-4bc2-4575-bb60-815ea6ae6a84-kube-api-access-sjprj" (OuterVolumeSpecName: "kube-api-access-sjprj") pod "f344130d-4bc2-4575-bb60-815ea6ae6a84" (UID: "f344130d-4bc2-4575-bb60-815ea6ae6a84"). InnerVolumeSpecName "kube-api-access-sjprj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.246237 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-kube-api-access-9vvfc" (OuterVolumeSpecName: "kube-api-access-9vvfc") pod "a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" (UID: "a5e4dfeb-181d-43c7-89bb-3c828fb5ce56"). InnerVolumeSpecName "kube-api-access-9vvfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.247232 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/454e49f6-b054-4901-bd18-461d7e9cd89e-kube-api-access-rnqqf" (OuterVolumeSpecName: "kube-api-access-rnqqf") pod "454e49f6-b054-4901-bd18-461d7e9cd89e" (UID: "454e49f6-b054-4901-bd18-461d7e9cd89e"). InnerVolumeSpecName "kube-api-access-rnqqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.262998 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f344130d-4bc2-4575-bb60-815ea6ae6a84" (UID: "f344130d-4bc2-4575-bb60-815ea6ae6a84"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.264919 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bcc59fe2-e410-4646-be11-560d0f45938b" (UID: "bcc59fe2-e410-4646-be11-560d0f45938b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342879 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342917 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342928 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342937 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc59fe2-e410-4646-be11-560d0f45938b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342946 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mnvn\" (UniqueName: \"kubernetes.io/projected/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-kube-api-access-7mnvn\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342956 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vvfc\" (UniqueName: \"kubernetes.io/projected/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-kube-api-access-9vvfc\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342963 4954 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342971 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f344130d-4bc2-4575-bb60-815ea6ae6a84-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342979 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnqqf\" (UniqueName: \"kubernetes.io/projected/454e49f6-b054-4901-bd18-461d7e9cd89e-kube-api-access-rnqqf\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342986 4954 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/454e49f6-b054-4901-bd18-461d7e9cd89e-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.342997 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjprj\" (UniqueName: \"kubernetes.io/projected/f344130d-4bc2-4575-bb60-815ea6ae6a84-kube-api-access-sjprj\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.343005 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjj6v\" (UniqueName: \"kubernetes.io/projected/bcc59fe2-e410-4646-be11-560d0f45938b-kube-api-access-sjj6v\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.343013 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.364328 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" (UID: "a5e4dfeb-181d-43c7-89bb-3c828fb5ce56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.371987 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0dc2999-1020-4b4b-96ae-87ba32e34ad8" (UID: "d0dc2999-1020-4b4b-96ae-87ba32e34ad8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.444150 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.444182 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0dc2999-1020-4b4b-96ae-87ba32e34ad8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.465453 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.539033 4954 generic.go:334] "Generic (PLEG): container finished" podID="454e49f6-b054-4901-bd18-461d7e9cd89e" containerID="5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13" exitCode=0 Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.539103 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" event={"ID":"454e49f6-b054-4901-bd18-461d7e9cd89e","Type":"ContainerDied","Data":"5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13"} Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.539130 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" event={"ID":"454e49f6-b054-4901-bd18-461d7e9cd89e","Type":"ContainerDied","Data":"4e238e14ac372d025d9b80ec7ad745692b7beb1abacb7e2aefa46d419d3e5184"} Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.539147 4954 scope.go:117] "RemoveContainer" containerID="5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.539246 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8jkvw" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.544135 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9xz9" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.544118 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9xz9" event={"ID":"f344130d-4bc2-4575-bb60-815ea6ae6a84","Type":"ContainerDied","Data":"2cd1325e49b8a5e2d271d52cd787123c775adf449b5ac0b59c57a02b9bd05874"} Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.549474 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjxbl" event={"ID":"c9ebbd26-bcdc-4db3-b345-a5715ac59a82","Type":"ContainerDied","Data":"ded987bb7da373d6123e13a47a53040e806b2aaf10006f0503aacf00428712ab"} Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.549558 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjxbl" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.554868 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wpj95" event={"ID":"a5e4dfeb-181d-43c7-89bb-3c828fb5ce56","Type":"ContainerDied","Data":"cc5f1f84b1d876464d860fe9634e3634e3c15a18001e682c9fef52b1ee4970be"} Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.554890 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wpj95" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.559246 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nlf5c" event={"ID":"271559e2-f374-4aeb-a3f7-6b039afc7e8f","Type":"ContainerDied","Data":"b28cf9b500bb266e35dbafe339f173310868e786e6f62ba93da51fc5486d8342"} Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.559455 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nlf5c" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.563354 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7mwfq" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.564351 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7mwfq" event={"ID":"bcc59fe2-e410-4646-be11-560d0f45938b","Type":"ContainerDied","Data":"bcaebf94cf10f78527abfeb2876fc1f449ff6538cc4ca8f43e4a24a704296945"} Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.569086 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vfz9w" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.569233 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vfz9w" event={"ID":"d0dc2999-1020-4b4b-96ae-87ba32e34ad8","Type":"ContainerDied","Data":"80bc71a0bf53a88cfdada5e765b97516cfd9ba88ebeeef6cb451e3c7a4c8f87b"} Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.572138 4954 scope.go:117] "RemoveContainer" containerID="5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13" Dec 09 17:01:58 crc kubenswrapper[4954]: E1209 17:01:58.573800 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13\": container with ID starting with 5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13 not found: ID does not exist" containerID="5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.573854 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13"} err="failed to get container status \"5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13\": rpc error: code = NotFound desc = could not find container \"5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13\": container with ID starting with 5f6ab86ae3afc91485b7f290b9690325b52f9e6f1ec9c781b88ea09fe69a9f13 not found: ID does not exist" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.573890 4954 scope.go:117] "RemoveContainer" containerID="975ac2f9c9073ce03cdb134f155cb901b644766788af76e2c3348b40c59adfed" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.576492 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j57d6" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.576678 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57d6" event={"ID":"f28a7275-4712-4dc9-89a3-16a77bba72d1","Type":"ContainerDied","Data":"8dcea5cde4292cf3e3b5f0ecf7804687718f4456131e9b2d7bbf191afba8827b"} Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.580376 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-snfrc" event={"ID":"1df4d5f7-03d7-4b91-b949-b555dae7ce2b","Type":"ContainerDied","Data":"66b3195cf75c18ea7476f3b74fdc85ae98b35d7058605fe6b7e367dca7ebae85"} Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.580440 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-snfrc" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.589329 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8jkvw"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.593571 4954 scope.go:117] "RemoveContainer" containerID="7cb820c86004e8b47d2826694a262ee6143c5f7066ee1facb495127620039a5b" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.598409 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8jkvw"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.620844 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jjxbl"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.631313 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jjxbl"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.636348 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7mwfq"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.643453 4954 scope.go:117] "RemoveContainer" containerID="78a8cfdf2f09063c9e057115de3ecad74edf3bb4dd5d3017be89c2c80dd6f243" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.643542 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7mwfq"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.653416 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nlf5c"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.661590 4954 scope.go:117] "RemoveContainer" containerID="b67d954d432fdaefd19cddd78c8dc54206aa1618d67cb9535fc075be133e30b2" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.669178 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nlf5c"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.678609 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-snfrc"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.687778 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-snfrc"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.689149 4954 scope.go:117] "RemoveContainer" containerID="cf55d9085bdaf7b9669d4fd8b2b8146e506b22cc4c443935b95201e5c900f380" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.692441 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j57d6"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.695312 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j57d6"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.698222 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9xz9"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.701483 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9xz9"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.705926 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wpj95"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.710994 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wpj95"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.714457 4954 scope.go:117] "RemoveContainer" containerID="bb637abb2df47f522f090fb6b4550c185f4cc3f38e406089a64aa08526d9a168" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.718161 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vfz9w"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.722651 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vfz9w"] Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.729164 4954 scope.go:117] "RemoveContainer" containerID="2b2d50b62b94bdd59bbb69d255d656e435b51c289d0ed62839023139383a3879" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.744754 4954 scope.go:117] "RemoveContainer" containerID="55a050768b735b0c202117de2c9a3f0b7f1fec8a58d5438b40597da9a6adc931" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.761791 4954 scope.go:117] "RemoveContainer" containerID="32a2781665742e8ec23731658a2f67bd96dcd24604b3a613a3ce82a2d575a87e" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.775107 4954 scope.go:117] "RemoveContainer" containerID="76b76e8114b08f3a9678e802e06fa776197be9e1d949be861a4bbe1b9688c346" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.780682 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.794662 4954 scope.go:117] "RemoveContainer" containerID="968568e4eee1faa4569a82635ca2e523923a4b9f547c7642c467122a91ef6cf1" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.809946 4954 scope.go:117] "RemoveContainer" containerID="10bf1fccc4e2742455309e9f198f00a2e607aa3a3def637556a839c0e59ade8c" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.823733 4954 scope.go:117] "RemoveContainer" containerID="e5fbba9ad18b1bb6fef18a9cf8bd59c22fa0955b79ee6e8bfb29ae7dfd46e8df" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.837497 4954 scope.go:117] "RemoveContainer" containerID="31768f2a27b179b400d6e8a37d899d36900883a9451a69d53d4ff72cee5307bf" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.851455 4954 scope.go:117] "RemoveContainer" containerID="d60e56fbf6d62aa1fbc2d97a997ebb9f713fa11ff65c52000c0826f71ed9655c" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.865564 4954 scope.go:117] "RemoveContainer" containerID="d3024a2ca08374fb704f4e2ffd8dce33c8ac5bf54e506b3c8e69e4edf115b0b8" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.890594 4954 scope.go:117] "RemoveContainer" containerID="b25bdc1a0ac311120b1b070d863514e204cf6251afe24724215fe7a7dff45d43" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.910115 4954 scope.go:117] "RemoveContainer" containerID="7b11d65a4c6d573ca717ed9eddb0f0fcaa8c176aa01652ee0d728205f85892e4" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.926211 4954 scope.go:117] "RemoveContainer" containerID="474a91cd7207e014c2d4d28ffbb23669f09afd2ee82fd6fcde0d2bfafbe7845d" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.939519 4954 scope.go:117] "RemoveContainer" containerID="e0446210a7a4e5f9689cf10fb03e60e90f775ea68962a4cff3896e743ad926f0" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.957049 4954 scope.go:117] "RemoveContainer" containerID="44412eee609a047930cd88f51f8174332c91aba076269be6583ddfc271cf63ed" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.972236 4954 scope.go:117] "RemoveContainer" containerID="62ba3752232989ae967c4086a4ac4d99e2ef36ed7c7dcd22835f3ba01353318b" Dec 09 17:01:58 crc kubenswrapper[4954]: I1209 17:01:58.986310 4954 scope.go:117] "RemoveContainer" containerID="4dfa72dd47b8b7a5c5f2643b19ff9e2b51912bf2e84293a6f40709a76ee57a9e" Dec 09 17:01:59 crc kubenswrapper[4954]: I1209 17:01:59.003509 4954 scope.go:117] "RemoveContainer" containerID="3f6fd90edf925a9aae6462e32442f363cc5fbf2f72578b67612d6f3c119a7035" Dec 09 17:02:00 crc kubenswrapper[4954]: I1209 17:02:00.135773 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" path="/var/lib/kubelet/pods/1df4d5f7-03d7-4b91-b949-b555dae7ce2b/volumes" Dec 09 17:02:00 crc kubenswrapper[4954]: I1209 17:02:00.137961 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" path="/var/lib/kubelet/pods/271559e2-f374-4aeb-a3f7-6b039afc7e8f/volumes" Dec 09 17:02:00 crc kubenswrapper[4954]: I1209 17:02:00.139340 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="454e49f6-b054-4901-bd18-461d7e9cd89e" path="/var/lib/kubelet/pods/454e49f6-b054-4901-bd18-461d7e9cd89e/volumes" Dec 09 17:02:00 crc kubenswrapper[4954]: I1209 17:02:00.143789 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" path="/var/lib/kubelet/pods/a5e4dfeb-181d-43c7-89bb-3c828fb5ce56/volumes" Dec 09 17:02:00 crc kubenswrapper[4954]: I1209 17:02:00.145091 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" path="/var/lib/kubelet/pods/bcc59fe2-e410-4646-be11-560d0f45938b/volumes" Dec 09 17:02:00 crc kubenswrapper[4954]: I1209 17:02:00.147504 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" path="/var/lib/kubelet/pods/c9ebbd26-bcdc-4db3-b345-a5715ac59a82/volumes" Dec 09 17:02:00 crc kubenswrapper[4954]: I1209 17:02:00.149005 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" path="/var/lib/kubelet/pods/d0dc2999-1020-4b4b-96ae-87ba32e34ad8/volumes" Dec 09 17:02:00 crc kubenswrapper[4954]: I1209 17:02:00.150338 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" path="/var/lib/kubelet/pods/f28a7275-4712-4dc9-89a3-16a77bba72d1/volumes" Dec 09 17:02:00 crc kubenswrapper[4954]: I1209 17:02:00.152642 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" path="/var/lib/kubelet/pods/f344130d-4bc2-4575-bb60-815ea6ae6a84/volumes" Dec 09 17:02:03 crc kubenswrapper[4954]: I1209 17:02:03.155965 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 17:02:06 crc kubenswrapper[4954]: I1209 17:02:06.079820 4954 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 17:02:06 crc kubenswrapper[4954]: I1209 17:02:06.080698 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34" gracePeriod=5 Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.647476 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.648183 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.670425 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.670472 4954 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34" exitCode=137 Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.670514 4954 scope.go:117] "RemoveContainer" containerID="688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.670526 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.686287 4954 scope.go:117] "RemoveContainer" containerID="688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34" Dec 09 17:02:11 crc kubenswrapper[4954]: E1209 17:02:11.686646 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34\": container with ID starting with 688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34 not found: ID does not exist" containerID="688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.686678 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34"} err="failed to get container status \"688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34\": rpc error: code = NotFound desc = could not find container \"688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34\": container with ID starting with 688ea326a269a05642fe0d423f0274f1e25cd775d666a5774ed995cc1c745d34 not found: ID does not exist" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.813481 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.813556 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.813649 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.813675 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.813635 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.813670 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.813691 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.813701 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.813756 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.814151 4954 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.814163 4954 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.814171 4954 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.814179 4954 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.821642 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:02:11 crc kubenswrapper[4954]: I1209 17:02:11.915003 4954 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:12 crc kubenswrapper[4954]: I1209 17:02:12.127482 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 09 17:02:12 crc kubenswrapper[4954]: I1209 17:02:12.127787 4954 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 09 17:02:12 crc kubenswrapper[4954]: I1209 17:02:12.138678 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 17:02:12 crc kubenswrapper[4954]: I1209 17:02:12.138725 4954 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="040bc860-1fd3-4a8b-aa43-0f088117fdb3" Dec 09 17:02:12 crc kubenswrapper[4954]: I1209 17:02:12.142454 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 17:02:12 crc kubenswrapper[4954]: I1209 17:02:12.142516 4954 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="040bc860-1fd3-4a8b-aa43-0f088117fdb3" Dec 09 17:02:16 crc kubenswrapper[4954]: I1209 17:02:16.325141 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 09 17:02:17 crc kubenswrapper[4954]: I1209 17:02:17.705546 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 09 17:02:17 crc kubenswrapper[4954]: I1209 17:02:17.707495 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 09 17:02:17 crc kubenswrapper[4954]: I1209 17:02:17.707541 4954 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="0d96526802b83b303759cac50ad985dea04fd0a6d5b3f45b83d3be452a1d1a43" exitCode=137 Dec 09 17:02:17 crc kubenswrapper[4954]: I1209 17:02:17.707575 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"0d96526802b83b303759cac50ad985dea04fd0a6d5b3f45b83d3be452a1d1a43"} Dec 09 17:02:17 crc kubenswrapper[4954]: I1209 17:02:17.707748 4954 scope.go:117] "RemoveContainer" containerID="a5004cc4e50aecd85dd6f1e14fdeb209ee7f7f43afc1104f8da231d1f3d6a53f" Dec 09 17:02:18 crc kubenswrapper[4954]: I1209 17:02:18.716345 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 09 17:02:18 crc kubenswrapper[4954]: I1209 17:02:18.718130 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c843ef6787190336f1c04e20d21ad4f5825aa86f151474f830936d91777cd4a4"} Dec 09 17:02:19 crc kubenswrapper[4954]: I1209 17:02:19.219143 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745014 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-45kb2"] Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745564 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745580 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745610 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745618 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745627 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745635 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745646 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745653 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745661 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745667 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745678 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745685 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745692 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745699 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745708 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745716 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745725 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745732 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745741 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745748 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745755 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745762 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745771 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="454e49f6-b054-4901-bd18-461d7e9cd89e" containerName="marketplace-operator" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745781 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="454e49f6-b054-4901-bd18-461d7e9cd89e" containerName="marketplace-operator" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745791 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745798 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745807 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745814 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745822 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745830 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745843 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745852 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745862 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745870 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745880 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745886 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745897 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745903 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745913 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745921 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745928 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745935 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerName="extract-utilities" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745942 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" containerName="installer" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745948 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" containerName="installer" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745957 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745963 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745973 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745979 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.745987 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.745994 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.746003 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746009 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: E1209 17:02:22.746020 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746028 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerName="extract-content" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746153 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcc59fe2-e410-4646-be11-560d0f45938b" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746165 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f28a7275-4712-4dc9-89a3-16a77bba72d1" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746174 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746183 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="271559e2-f374-4aeb-a3f7-6b039afc7e8f" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746191 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5e4dfeb-181d-43c7-89bb-3c828fb5ce56" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746202 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce53dd1a-0beb-4c91-9422-8473bd7a9412" containerName="installer" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746209 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="454e49f6-b054-4901-bd18-461d7e9cd89e" containerName="marketplace-operator" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746217 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0dc2999-1020-4b4b-96ae-87ba32e34ad8" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746228 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df4d5f7-03d7-4b91-b949-b555dae7ce2b" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746237 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f344130d-4bc2-4575-bb60-815ea6ae6a84" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.746245 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9ebbd26-bcdc-4db3-b345-a5715ac59a82" containerName="registry-server" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.747135 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.750136 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.751230 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.755580 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.757230 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-45kb2"] Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.852252 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct2v2\" (UniqueName: \"kubernetes.io/projected/edc2fc95-fd10-469e-9e70-8368b75d1d82-kube-api-access-ct2v2\") pod \"certified-operators-45kb2\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.852496 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-utilities\") pod \"certified-operators-45kb2\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.852815 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-catalog-content\") pod \"certified-operators-45kb2\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.942961 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.954531 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-catalog-content\") pod \"certified-operators-45kb2\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.954714 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct2v2\" (UniqueName: \"kubernetes.io/projected/edc2fc95-fd10-469e-9e70-8368b75d1d82-kube-api-access-ct2v2\") pod \"certified-operators-45kb2\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.954755 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-utilities\") pod \"certified-operators-45kb2\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.955232 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-catalog-content\") pod \"certified-operators-45kb2\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.955352 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-utilities\") pod \"certified-operators-45kb2\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:22 crc kubenswrapper[4954]: I1209 17:02:22.976353 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct2v2\" (UniqueName: \"kubernetes.io/projected/edc2fc95-fd10-469e-9e70-8368b75d1d82-kube-api-access-ct2v2\") pod \"certified-operators-45kb2\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.064538 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.315993 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-45kb2"] Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.341857 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g4cpr"] Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.342950 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.347544 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.359357 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g4cpr"] Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.459897 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-catalog-content\") pod \"community-operators-g4cpr\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.460490 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-utilities\") pod \"community-operators-g4cpr\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.460536 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv5dh\" (UniqueName: \"kubernetes.io/projected/da66ed1d-0022-4970-bc4c-aa6a62df0755-kube-api-access-hv5dh\") pod \"community-operators-g4cpr\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.562733 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-utilities\") pod \"community-operators-g4cpr\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.562885 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv5dh\" (UniqueName: \"kubernetes.io/projected/da66ed1d-0022-4970-bc4c-aa6a62df0755-kube-api-access-hv5dh\") pod \"community-operators-g4cpr\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.562927 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-catalog-content\") pod \"community-operators-g4cpr\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.563697 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-catalog-content\") pod \"community-operators-g4cpr\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.563774 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-utilities\") pod \"community-operators-g4cpr\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.584436 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv5dh\" (UniqueName: \"kubernetes.io/projected/da66ed1d-0022-4970-bc4c-aa6a62df0755-kube-api-access-hv5dh\") pod \"community-operators-g4cpr\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.702509 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.759665 4954 generic.go:334] "Generic (PLEG): container finished" podID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerID="ff2f8a68fa23172a3d3184ea750e5a76bc4e281859d536e859426013b7503cc0" exitCode=0 Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.759713 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45kb2" event={"ID":"edc2fc95-fd10-469e-9e70-8368b75d1d82","Type":"ContainerDied","Data":"ff2f8a68fa23172a3d3184ea750e5a76bc4e281859d536e859426013b7503cc0"} Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.759743 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45kb2" event={"ID":"edc2fc95-fd10-469e-9e70-8368b75d1d82","Type":"ContainerStarted","Data":"8ccd1e4a1052f69b487e4679d4d6154d700e592b261ff1d058c0521a9c7fcf76"} Dec 09 17:02:23 crc kubenswrapper[4954]: I1209 17:02:23.940783 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g4cpr"] Dec 09 17:02:24 crc kubenswrapper[4954]: I1209 17:02:24.766076 4954 generic.go:334] "Generic (PLEG): container finished" podID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerID="2ba8d8af2a88d2314467f4d684e5e8381510d4990cdf2e34256ce7e0be366c1f" exitCode=0 Dec 09 17:02:24 crc kubenswrapper[4954]: I1209 17:02:24.766305 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4cpr" event={"ID":"da66ed1d-0022-4970-bc4c-aa6a62df0755","Type":"ContainerDied","Data":"2ba8d8af2a88d2314467f4d684e5e8381510d4990cdf2e34256ce7e0be366c1f"} Dec 09 17:02:24 crc kubenswrapper[4954]: I1209 17:02:24.766782 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4cpr" event={"ID":"da66ed1d-0022-4970-bc4c-aa6a62df0755","Type":"ContainerStarted","Data":"60289ebc00941a667f985407b5d50918c8910be7d5a67f895e353a9ea36001d5"} Dec 09 17:02:24 crc kubenswrapper[4954]: I1209 17:02:24.769085 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45kb2" event={"ID":"edc2fc95-fd10-469e-9e70-8368b75d1d82","Type":"ContainerStarted","Data":"ee355a22239b8c9839879a4b35a6354ba052d7c9447f43f3356fc47676aa3744"} Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.142836 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7x66v"] Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.144140 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.146849 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.154944 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7x66v"] Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.285162 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-catalog-content\") pod \"redhat-marketplace-7x66v\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.285703 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhrdr\" (UniqueName: \"kubernetes.io/projected/4084f661-2955-43ba-a2d4-e6389830d100-kube-api-access-rhrdr\") pod \"redhat-marketplace-7x66v\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.285764 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-utilities\") pod \"redhat-marketplace-7x66v\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.386480 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhrdr\" (UniqueName: \"kubernetes.io/projected/4084f661-2955-43ba-a2d4-e6389830d100-kube-api-access-rhrdr\") pod \"redhat-marketplace-7x66v\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.386533 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-utilities\") pod \"redhat-marketplace-7x66v\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.386560 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-catalog-content\") pod \"redhat-marketplace-7x66v\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.387037 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-catalog-content\") pod \"redhat-marketplace-7x66v\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.387134 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-utilities\") pod \"redhat-marketplace-7x66v\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.413108 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhrdr\" (UniqueName: \"kubernetes.io/projected/4084f661-2955-43ba-a2d4-e6389830d100-kube-api-access-rhrdr\") pod \"redhat-marketplace-7x66v\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.461119 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.675472 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7x66v"] Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.754290 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q8lch"] Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.756318 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.763273 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.764870 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8lch"] Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.777577 4954 generic.go:334] "Generic (PLEG): container finished" podID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerID="ee355a22239b8c9839879a4b35a6354ba052d7c9447f43f3356fc47676aa3744" exitCode=0 Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.777665 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45kb2" event={"ID":"edc2fc95-fd10-469e-9e70-8368b75d1d82","Type":"ContainerDied","Data":"ee355a22239b8c9839879a4b35a6354ba052d7c9447f43f3356fc47676aa3744"} Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.779170 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7x66v" event={"ID":"4084f661-2955-43ba-a2d4-e6389830d100","Type":"ContainerStarted","Data":"f91cfe648e0d4d35425d69aaebc9be4a68610ed3503f81a6556322492d38f750"} Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.896949 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lhwm\" (UniqueName: \"kubernetes.io/projected/30f323d5-cbe3-4209-a708-a2e277eab190-kube-api-access-2lhwm\") pod \"redhat-operators-q8lch\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.897006 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-catalog-content\") pod \"redhat-operators-q8lch\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.897146 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-utilities\") pod \"redhat-operators-q8lch\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.998712 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lhwm\" (UniqueName: \"kubernetes.io/projected/30f323d5-cbe3-4209-a708-a2e277eab190-kube-api-access-2lhwm\") pod \"redhat-operators-q8lch\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.998857 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-catalog-content\") pod \"redhat-operators-q8lch\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.999055 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-utilities\") pod \"redhat-operators-q8lch\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.999273 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-catalog-content\") pod \"redhat-operators-q8lch\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:25 crc kubenswrapper[4954]: I1209 17:02:25.999518 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-utilities\") pod \"redhat-operators-q8lch\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:26 crc kubenswrapper[4954]: I1209 17:02:26.018682 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lhwm\" (UniqueName: \"kubernetes.io/projected/30f323d5-cbe3-4209-a708-a2e277eab190-kube-api-access-2lhwm\") pod \"redhat-operators-q8lch\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:26 crc kubenswrapper[4954]: I1209 17:02:26.077182 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:26 crc kubenswrapper[4954]: I1209 17:02:26.267928 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8lch"] Dec 09 17:02:26 crc kubenswrapper[4954]: W1209 17:02:26.274036 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30f323d5_cbe3_4209_a708_a2e277eab190.slice/crio-6b7b8c08980f4a18b9830b023a1a4630d69f0ac5b2becf62a2d4b5f6ba375fb8 WatchSource:0}: Error finding container 6b7b8c08980f4a18b9830b023a1a4630d69f0ac5b2becf62a2d4b5f6ba375fb8: Status 404 returned error can't find the container with id 6b7b8c08980f4a18b9830b023a1a4630d69f0ac5b2becf62a2d4b5f6ba375fb8 Dec 09 17:02:26 crc kubenswrapper[4954]: I1209 17:02:26.787392 4954 generic.go:334] "Generic (PLEG): container finished" podID="30f323d5-cbe3-4209-a708-a2e277eab190" containerID="83436d1245c55f2a5637d35967cf5b613ca849dce49eb02613142b119b520899" exitCode=0 Dec 09 17:02:26 crc kubenswrapper[4954]: I1209 17:02:26.787502 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8lch" event={"ID":"30f323d5-cbe3-4209-a708-a2e277eab190","Type":"ContainerDied","Data":"83436d1245c55f2a5637d35967cf5b613ca849dce49eb02613142b119b520899"} Dec 09 17:02:26 crc kubenswrapper[4954]: I1209 17:02:26.788030 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8lch" event={"ID":"30f323d5-cbe3-4209-a708-a2e277eab190","Type":"ContainerStarted","Data":"6b7b8c08980f4a18b9830b023a1a4630d69f0ac5b2becf62a2d4b5f6ba375fb8"} Dec 09 17:02:26 crc kubenswrapper[4954]: I1209 17:02:26.790679 4954 generic.go:334] "Generic (PLEG): container finished" podID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerID="01251ec2969f4ee6b355ba129fd60509016f6c24fe5f9f7cbf928ce6575d30c3" exitCode=0 Dec 09 17:02:26 crc kubenswrapper[4954]: I1209 17:02:26.790838 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4cpr" event={"ID":"da66ed1d-0022-4970-bc4c-aa6a62df0755","Type":"ContainerDied","Data":"01251ec2969f4ee6b355ba129fd60509016f6c24fe5f9f7cbf928ce6575d30c3"} Dec 09 17:02:26 crc kubenswrapper[4954]: I1209 17:02:26.793644 4954 generic.go:334] "Generic (PLEG): container finished" podID="4084f661-2955-43ba-a2d4-e6389830d100" containerID="109e5810b1a657ab4402f83b2e1eaf0cc9eb8ba389364f03b5eba3c680f0b95c" exitCode=0 Dec 09 17:02:26 crc kubenswrapper[4954]: I1209 17:02:26.793697 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7x66v" event={"ID":"4084f661-2955-43ba-a2d4-e6389830d100","Type":"ContainerDied","Data":"109e5810b1a657ab4402f83b2e1eaf0cc9eb8ba389364f03b5eba3c680f0b95c"} Dec 09 17:02:27 crc kubenswrapper[4954]: I1209 17:02:27.554045 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 17:02:27 crc kubenswrapper[4954]: I1209 17:02:27.558291 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 17:02:27 crc kubenswrapper[4954]: I1209 17:02:27.803377 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45kb2" event={"ID":"edc2fc95-fd10-469e-9e70-8368b75d1d82","Type":"ContainerStarted","Data":"72df4cc2d6b4b84c5cb6dfc9d53e848945487680f3f16efd084b2850dffea7f9"} Dec 09 17:02:27 crc kubenswrapper[4954]: I1209 17:02:27.808572 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8lch" event={"ID":"30f323d5-cbe3-4209-a708-a2e277eab190","Type":"ContainerStarted","Data":"be44a0e2910f1dd296bf694436bb1194a03295ced9176724b66b2d24fabacb6a"} Dec 09 17:02:27 crc kubenswrapper[4954]: I1209 17:02:27.814720 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 17:02:27 crc kubenswrapper[4954]: I1209 17:02:27.825123 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-45kb2" podStartSLOduration=2.313214471 podStartE2EDuration="5.825100118s" podCreationTimestamp="2025-12-09 17:02:22 +0000 UTC" firstStartedPulling="2025-12-09 17:02:23.761740864 +0000 UTC m=+340.149914684" lastFinishedPulling="2025-12-09 17:02:27.273626511 +0000 UTC m=+343.661800331" observedRunningTime="2025-12-09 17:02:27.824296342 +0000 UTC m=+344.212470172" watchObservedRunningTime="2025-12-09 17:02:27.825100118 +0000 UTC m=+344.213273938" Dec 09 17:02:28 crc kubenswrapper[4954]: I1209 17:02:28.814972 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4cpr" event={"ID":"da66ed1d-0022-4970-bc4c-aa6a62df0755","Type":"ContainerStarted","Data":"697f4a455885134b6769d2283d6a950e72e67e1cf3a4ddab3183dbeeabab9213"} Dec 09 17:02:28 crc kubenswrapper[4954]: I1209 17:02:28.817098 4954 generic.go:334] "Generic (PLEG): container finished" podID="4084f661-2955-43ba-a2d4-e6389830d100" containerID="d106d7667edbb1b61e291ea928a889a5ac8bb54f728fc55c64fd660a671061ba" exitCode=0 Dec 09 17:02:28 crc kubenswrapper[4954]: I1209 17:02:28.817149 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7x66v" event={"ID":"4084f661-2955-43ba-a2d4-e6389830d100","Type":"ContainerDied","Data":"d106d7667edbb1b61e291ea928a889a5ac8bb54f728fc55c64fd660a671061ba"} Dec 09 17:02:28 crc kubenswrapper[4954]: I1209 17:02:28.819992 4954 generic.go:334] "Generic (PLEG): container finished" podID="30f323d5-cbe3-4209-a708-a2e277eab190" containerID="be44a0e2910f1dd296bf694436bb1194a03295ced9176724b66b2d24fabacb6a" exitCode=0 Dec 09 17:02:28 crc kubenswrapper[4954]: I1209 17:02:28.820072 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8lch" event={"ID":"30f323d5-cbe3-4209-a708-a2e277eab190","Type":"ContainerDied","Data":"be44a0e2910f1dd296bf694436bb1194a03295ced9176724b66b2d24fabacb6a"} Dec 09 17:02:28 crc kubenswrapper[4954]: I1209 17:02:28.840055 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g4cpr" podStartSLOduration=3.049080655 podStartE2EDuration="5.84003333s" podCreationTimestamp="2025-12-09 17:02:23 +0000 UTC" firstStartedPulling="2025-12-09 17:02:24.768219034 +0000 UTC m=+341.156392854" lastFinishedPulling="2025-12-09 17:02:27.559171709 +0000 UTC m=+343.947345529" observedRunningTime="2025-12-09 17:02:28.837267861 +0000 UTC m=+345.225441681" watchObservedRunningTime="2025-12-09 17:02:28.84003333 +0000 UTC m=+345.228207150" Dec 09 17:02:30 crc kubenswrapper[4954]: I1209 17:02:30.833623 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8lch" event={"ID":"30f323d5-cbe3-4209-a708-a2e277eab190","Type":"ContainerStarted","Data":"9289442f940546cbcd543822d629ae49e411d9d8eb95f4dde17e9bebf5ff4156"} Dec 09 17:02:30 crc kubenswrapper[4954]: I1209 17:02:30.836607 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7x66v" event={"ID":"4084f661-2955-43ba-a2d4-e6389830d100","Type":"ContainerStarted","Data":"35cdf07ce6733998fff22a56eb25f59a8ba9b435366fa01e6aac64db13999372"} Dec 09 17:02:30 crc kubenswrapper[4954]: I1209 17:02:30.868271 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q8lch" podStartSLOduration=3.07708816 podStartE2EDuration="5.868250831s" podCreationTimestamp="2025-12-09 17:02:25 +0000 UTC" firstStartedPulling="2025-12-09 17:02:26.78971942 +0000 UTC m=+343.177893240" lastFinishedPulling="2025-12-09 17:02:29.580882091 +0000 UTC m=+345.969055911" observedRunningTime="2025-12-09 17:02:30.862949521 +0000 UTC m=+347.251123351" watchObservedRunningTime="2025-12-09 17:02:30.868250831 +0000 UTC m=+347.256424651" Dec 09 17:02:30 crc kubenswrapper[4954]: I1209 17:02:30.889587 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7x66v" podStartSLOduration=3.093253538 podStartE2EDuration="5.889565985s" podCreationTimestamp="2025-12-09 17:02:25 +0000 UTC" firstStartedPulling="2025-12-09 17:02:26.795300669 +0000 UTC m=+343.183474479" lastFinishedPulling="2025-12-09 17:02:29.591613106 +0000 UTC m=+345.979786926" observedRunningTime="2025-12-09 17:02:30.888268053 +0000 UTC m=+347.276441893" watchObservedRunningTime="2025-12-09 17:02:30.889565985 +0000 UTC m=+347.277739805" Dec 09 17:02:33 crc kubenswrapper[4954]: I1209 17:02:33.065784 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:33 crc kubenswrapper[4954]: I1209 17:02:33.066143 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:33 crc kubenswrapper[4954]: I1209 17:02:33.109475 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:33 crc kubenswrapper[4954]: I1209 17:02:33.703576 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:33 crc kubenswrapper[4954]: I1209 17:02:33.703704 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:33 crc kubenswrapper[4954]: I1209 17:02:33.750359 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:33 crc kubenswrapper[4954]: I1209 17:02:33.890755 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:02:33 crc kubenswrapper[4954]: I1209 17:02:33.893972 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:02:35 crc kubenswrapper[4954]: I1209 17:02:35.462272 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:35 crc kubenswrapper[4954]: I1209 17:02:35.462705 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:35 crc kubenswrapper[4954]: I1209 17:02:35.502088 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:35 crc kubenswrapper[4954]: I1209 17:02:35.901091 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:02:36 crc kubenswrapper[4954]: I1209 17:02:36.077569 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:36 crc kubenswrapper[4954]: I1209 17:02:36.077676 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:36 crc kubenswrapper[4954]: I1209 17:02:36.127558 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:36 crc kubenswrapper[4954]: I1209 17:02:36.907899 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:02:43 crc kubenswrapper[4954]: I1209 17:02:43.756364 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:02:43 crc kubenswrapper[4954]: I1209 17:02:43.757243 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.108546 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v98r2"] Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.109774 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.117220 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.117526 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.123626 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.136373 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v98r2"] Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.171996 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bg5hn"] Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.178259 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.185643 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mnc4f"] Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.185942 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" podUID="1cb535ff-3daa-418e-929e-99664eea736b" containerName="controller-manager" containerID="cri-o://121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4" gracePeriod=30 Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.200679 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp"] Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.201042 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" podUID="6a770208-d33d-48d7-b70c-be193682fb72" containerName="route-controller-manager" containerID="cri-o://0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823" gracePeriod=30 Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.219801 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bg5hn"] Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281468 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhs5q\" (UniqueName: \"kubernetes.io/projected/6ff181e0-303d-4c55-9fc5-8c80561714ca-kube-api-access-dhs5q\") pod \"marketplace-operator-79b997595-v98r2\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281542 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v98r2\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281566 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjjrt\" (UniqueName: \"kubernetes.io/projected/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-kube-api-access-vjjrt\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281610 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-bound-sa-token\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281654 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281691 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281728 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281750 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-trusted-ca\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281783 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-registry-tls\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281814 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v98r2\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.281848 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-registry-certificates\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.336201 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.383191 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v98r2\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.383255 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-registry-certificates\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.383310 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhs5q\" (UniqueName: \"kubernetes.io/projected/6ff181e0-303d-4c55-9fc5-8c80561714ca-kube-api-access-dhs5q\") pod \"marketplace-operator-79b997595-v98r2\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.383335 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v98r2\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.383361 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-bound-sa-token\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.383383 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjjrt\" (UniqueName: \"kubernetes.io/projected/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-kube-api-access-vjjrt\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.383433 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.383467 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.383492 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-trusted-ca\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.383542 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-registry-tls\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.384750 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-registry-certificates\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.385155 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.385852 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v98r2\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.386530 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-trusted-ca\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.404367 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-registry-tls\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.409875 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v98r2\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.414217 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.415192 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjjrt\" (UniqueName: \"kubernetes.io/projected/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-kube-api-access-vjjrt\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.422872 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhs5q\" (UniqueName: \"kubernetes.io/projected/6ff181e0-303d-4c55-9fc5-8c80561714ca-kube-api-access-dhs5q\") pod \"marketplace-operator-79b997595-v98r2\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.428171 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c55383e9-5bdc-4404-802c-f7ad32a0ebe5-bound-sa-token\") pod \"image-registry-66df7c8f76-bg5hn\" (UID: \"c55383e9-5bdc-4404-802c-f7ad32a0ebe5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.435910 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.548655 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.705875 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.759649 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.790683 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-config\") pod \"6a770208-d33d-48d7-b70c-be193682fb72\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.790726 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-client-ca\") pod \"6a770208-d33d-48d7-b70c-be193682fb72\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.790798 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a770208-d33d-48d7-b70c-be193682fb72-serving-cert\") pod \"6a770208-d33d-48d7-b70c-be193682fb72\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.790852 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s6bz\" (UniqueName: \"kubernetes.io/projected/6a770208-d33d-48d7-b70c-be193682fb72-kube-api-access-6s6bz\") pod \"6a770208-d33d-48d7-b70c-be193682fb72\" (UID: \"6a770208-d33d-48d7-b70c-be193682fb72\") " Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.793094 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-client-ca" (OuterVolumeSpecName: "client-ca") pod "6a770208-d33d-48d7-b70c-be193682fb72" (UID: "6a770208-d33d-48d7-b70c-be193682fb72"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.793209 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-config" (OuterVolumeSpecName: "config") pod "6a770208-d33d-48d7-b70c-be193682fb72" (UID: "6a770208-d33d-48d7-b70c-be193682fb72"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.796201 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a770208-d33d-48d7-b70c-be193682fb72-kube-api-access-6s6bz" (OuterVolumeSpecName: "kube-api-access-6s6bz") pod "6a770208-d33d-48d7-b70c-be193682fb72" (UID: "6a770208-d33d-48d7-b70c-be193682fb72"). InnerVolumeSpecName "kube-api-access-6s6bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.796999 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a770208-d33d-48d7-b70c-be193682fb72-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6a770208-d33d-48d7-b70c-be193682fb72" (UID: "6a770208-d33d-48d7-b70c-be193682fb72"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.892093 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-client-ca\") pod \"1cb535ff-3daa-418e-929e-99664eea736b\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.892190 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cb535ff-3daa-418e-929e-99664eea736b-serving-cert\") pod \"1cb535ff-3daa-418e-929e-99664eea736b\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.892233 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7n2c\" (UniqueName: \"kubernetes.io/projected/1cb535ff-3daa-418e-929e-99664eea736b-kube-api-access-r7n2c\") pod \"1cb535ff-3daa-418e-929e-99664eea736b\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.892270 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-proxy-ca-bundles\") pod \"1cb535ff-3daa-418e-929e-99664eea736b\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.892365 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-config\") pod \"1cb535ff-3daa-418e-929e-99664eea736b\" (UID: \"1cb535ff-3daa-418e-929e-99664eea736b\") " Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.892728 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a770208-d33d-48d7-b70c-be193682fb72-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.892751 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s6bz\" (UniqueName: \"kubernetes.io/projected/6a770208-d33d-48d7-b70c-be193682fb72-kube-api-access-6s6bz\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.892764 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.892778 4954 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a770208-d33d-48d7-b70c-be193682fb72-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.893291 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-client-ca" (OuterVolumeSpecName: "client-ca") pod "1cb535ff-3daa-418e-929e-99664eea736b" (UID: "1cb535ff-3daa-418e-929e-99664eea736b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.893311 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1cb535ff-3daa-418e-929e-99664eea736b" (UID: "1cb535ff-3daa-418e-929e-99664eea736b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.893515 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-config" (OuterVolumeSpecName: "config") pod "1cb535ff-3daa-418e-929e-99664eea736b" (UID: "1cb535ff-3daa-418e-929e-99664eea736b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.896804 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cb535ff-3daa-418e-929e-99664eea736b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1cb535ff-3daa-418e-929e-99664eea736b" (UID: "1cb535ff-3daa-418e-929e-99664eea736b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.897001 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cb535ff-3daa-418e-929e-99664eea736b-kube-api-access-r7n2c" (OuterVolumeSpecName: "kube-api-access-r7n2c") pod "1cb535ff-3daa-418e-929e-99664eea736b" (UID: "1cb535ff-3daa-418e-929e-99664eea736b"). InnerVolumeSpecName "kube-api-access-r7n2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.931456 4954 generic.go:334] "Generic (PLEG): container finished" podID="1cb535ff-3daa-418e-929e-99664eea736b" containerID="121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4" exitCode=0 Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.931520 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.931555 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" event={"ID":"1cb535ff-3daa-418e-929e-99664eea736b","Type":"ContainerDied","Data":"121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4"} Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.931629 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mnc4f" event={"ID":"1cb535ff-3daa-418e-929e-99664eea736b","Type":"ContainerDied","Data":"37a701a282458902b44da1488e10efb91809ae98d999e159ae4431c94bcde645"} Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.931654 4954 scope.go:117] "RemoveContainer" containerID="121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.934080 4954 generic.go:334] "Generic (PLEG): container finished" podID="6a770208-d33d-48d7-b70c-be193682fb72" containerID="0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823" exitCode=0 Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.934149 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" event={"ID":"6a770208-d33d-48d7-b70c-be193682fb72","Type":"ContainerDied","Data":"0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823"} Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.934205 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" event={"ID":"6a770208-d33d-48d7-b70c-be193682fb72","Type":"ContainerDied","Data":"53f691fbd0badf79c0d8b31af8226fb73c095d16cfcd33b5b2df0d0762f0f6f3"} Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.934162 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.952706 4954 scope.go:117] "RemoveContainer" containerID="121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4" Dec 09 17:02:46 crc kubenswrapper[4954]: E1209 17:02:46.953479 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4\": container with ID starting with 121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4 not found: ID does not exist" containerID="121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.953546 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4"} err="failed to get container status \"121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4\": rpc error: code = NotFound desc = could not find container \"121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4\": container with ID starting with 121f344c744921d5810a825a9b7f2edae6f3db2756a675dc45512af19e5baae4 not found: ID does not exist" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.953581 4954 scope.go:117] "RemoveContainer" containerID="0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.966247 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mnc4f"] Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.971985 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mnc4f"] Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.972431 4954 scope.go:117] "RemoveContainer" containerID="0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823" Dec 09 17:02:46 crc kubenswrapper[4954]: E1209 17:02:46.981968 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823\": container with ID starting with 0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823 not found: ID does not exist" containerID="0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.982033 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823"} err="failed to get container status \"0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823\": rpc error: code = NotFound desc = could not find container \"0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823\": container with ID starting with 0d573587b572156427044bea72a50a422027d1354946a018b72798de29230823 not found: ID does not exist" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.994174 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cb535ff-3daa-418e-929e-99664eea736b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.994210 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7n2c\" (UniqueName: \"kubernetes.io/projected/1cb535ff-3daa-418e-929e-99664eea736b-kube-api-access-r7n2c\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.994221 4954 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.994232 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.994241 4954 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1cb535ff-3daa-418e-929e-99664eea736b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 17:02:46 crc kubenswrapper[4954]: I1209 17:02:46.994968 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp"] Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.004164 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5cftp"] Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.094721 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bg5hn"] Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.103470 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v98r2"] Dec 09 17:02:47 crc kubenswrapper[4954]: W1209 17:02:47.108754 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc55383e9_5bdc_4404_802c_f7ad32a0ebe5.slice/crio-253c5710a33211ad9a168336a8402e7cc432e9372dc53ea2786f83bedeeb9c3a WatchSource:0}: Error finding container 253c5710a33211ad9a168336a8402e7cc432e9372dc53ea2786f83bedeeb9c3a: Status 404 returned error can't find the container with id 253c5710a33211ad9a168336a8402e7cc432e9372dc53ea2786f83bedeeb9c3a Dec 09 17:02:47 crc kubenswrapper[4954]: W1209 17:02:47.110525 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff181e0_303d_4c55_9fc5_8c80561714ca.slice/crio-e91f29c426f2f85f03f8352dfdf3843dde12af4e3d1d0cdc91538a69c3c97044 WatchSource:0}: Error finding container e91f29c426f2f85f03f8352dfdf3843dde12af4e3d1d0cdc91538a69c3c97044: Status 404 returned error can't find the container with id e91f29c426f2f85f03f8352dfdf3843dde12af4e3d1d0cdc91538a69c3c97044 Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.289289 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj"] Dec 09 17:02:47 crc kubenswrapper[4954]: E1209 17:02:47.289549 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a770208-d33d-48d7-b70c-be193682fb72" containerName="route-controller-manager" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.289570 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a770208-d33d-48d7-b70c-be193682fb72" containerName="route-controller-manager" Dec 09 17:02:47 crc kubenswrapper[4954]: E1209 17:02:47.289583 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cb535ff-3daa-418e-929e-99664eea736b" containerName="controller-manager" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.289622 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cb535ff-3daa-418e-929e-99664eea736b" containerName="controller-manager" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.289743 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cb535ff-3daa-418e-929e-99664eea736b" containerName="controller-manager" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.289760 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a770208-d33d-48d7-b70c-be193682fb72" containerName="route-controller-manager" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.290132 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.292352 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.293299 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.293341 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.293587 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.293803 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.296297 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.302813 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj"] Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.399395 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-serving-cert\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.400013 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-client-ca\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.400112 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-config\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.400151 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc87f\" (UniqueName: \"kubernetes.io/projected/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-kube-api-access-gc87f\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.502139 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc87f\" (UniqueName: \"kubernetes.io/projected/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-kube-api-access-gc87f\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.502214 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-serving-cert\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.502257 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-client-ca\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.502345 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-config\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.503488 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-client-ca\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.503940 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-config\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.507075 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-serving-cert\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.519397 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc87f\" (UniqueName: \"kubernetes.io/projected/0098dd54-32ab-4b62-8d99-da8d5cfb7ad8-kube-api-access-gc87f\") pod \"route-controller-manager-544bb94698-rhrxj\" (UID: \"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8\") " pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.612095 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.836934 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj"] Dec 09 17:02:47 crc kubenswrapper[4954]: W1209 17:02:47.839725 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0098dd54_32ab_4b62_8d99_da8d5cfb7ad8.slice/crio-cab6c4624f174b1d78ae42b1e853f496c257ed32f9ccb9e51b021ee29238797f WatchSource:0}: Error finding container cab6c4624f174b1d78ae42b1e853f496c257ed32f9ccb9e51b021ee29238797f: Status 404 returned error can't find the container with id cab6c4624f174b1d78ae42b1e853f496c257ed32f9ccb9e51b021ee29238797f Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.940425 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" event={"ID":"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8","Type":"ContainerStarted","Data":"cab6c4624f174b1d78ae42b1e853f496c257ed32f9ccb9e51b021ee29238797f"} Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.942048 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" event={"ID":"c55383e9-5bdc-4404-802c-f7ad32a0ebe5","Type":"ContainerStarted","Data":"9ca83a9e62ce721f4a8e62ef6bfa3cbc28dfa10354b66c87f0b4ac0cede90dfb"} Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.942123 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" event={"ID":"c55383e9-5bdc-4404-802c-f7ad32a0ebe5","Type":"ContainerStarted","Data":"253c5710a33211ad9a168336a8402e7cc432e9372dc53ea2786f83bedeeb9c3a"} Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.942173 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.944310 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" event={"ID":"6ff181e0-303d-4c55-9fc5-8c80561714ca","Type":"ContainerStarted","Data":"33195b1264b7a3fb8bf839bf88ae5b95c78bae9ae678023058d9a2aee9eab833"} Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.944343 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" event={"ID":"6ff181e0-303d-4c55-9fc5-8c80561714ca","Type":"ContainerStarted","Data":"e91f29c426f2f85f03f8352dfdf3843dde12af4e3d1d0cdc91538a69c3c97044"} Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.944532 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.949561 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.988254 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" podStartSLOduration=1.988234971 podStartE2EDuration="1.988234971s" podCreationTimestamp="2025-12-09 17:02:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:02:47.98601649 +0000 UTC m=+364.374190330" watchObservedRunningTime="2025-12-09 17:02:47.988234971 +0000 UTC m=+364.376408781" Dec 09 17:02:47 crc kubenswrapper[4954]: I1209 17:02:47.988841 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" podStartSLOduration=1.98883286 podStartE2EDuration="1.98883286s" podCreationTimestamp="2025-12-09 17:02:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:02:47.97013265 +0000 UTC m=+364.358306470" watchObservedRunningTime="2025-12-09 17:02:47.98883286 +0000 UTC m=+364.377006680" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.127727 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cb535ff-3daa-418e-929e-99664eea736b" path="/var/lib/kubelet/pods/1cb535ff-3daa-418e-929e-99664eea736b/volumes" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.128644 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a770208-d33d-48d7-b70c-be193682fb72" path="/var/lib/kubelet/pods/6a770208-d33d-48d7-b70c-be193682fb72/volumes" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.292339 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-868f679974-cxc7g"] Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.293407 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.297685 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.297926 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.298726 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.298972 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.299279 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.299489 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.310936 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.311360 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-868f679974-cxc7g"] Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.416000 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-serving-cert\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.416069 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-client-ca\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.416099 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-proxy-ca-bundles\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.416131 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-config\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.416192 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hjvc\" (UniqueName: \"kubernetes.io/projected/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-kube-api-access-5hjvc\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.517432 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-client-ca\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.517508 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-proxy-ca-bundles\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.517542 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-config\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.517611 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hjvc\" (UniqueName: \"kubernetes.io/projected/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-kube-api-access-5hjvc\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.517650 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-serving-cert\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.519871 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-client-ca\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.519944 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-proxy-ca-bundles\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.520363 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-config\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.524838 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-serving-cert\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.537394 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hjvc\" (UniqueName: \"kubernetes.io/projected/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-kube-api-access-5hjvc\") pod \"controller-manager-868f679974-cxc7g\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.613679 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.819240 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-868f679974-cxc7g"] Dec 09 17:02:48 crc kubenswrapper[4954]: W1209 17:02:48.822497 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod692c8d95_4a19_48c8_b70d_571ffbb1ca8e.slice/crio-d4a2cb14b630c139ec8b62781149735e04ea0beab94a09a69bcf87001c9e4a60 WatchSource:0}: Error finding container d4a2cb14b630c139ec8b62781149735e04ea0beab94a09a69bcf87001c9e4a60: Status 404 returned error can't find the container with id d4a2cb14b630c139ec8b62781149735e04ea0beab94a09a69bcf87001c9e4a60 Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.956318 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" event={"ID":"0098dd54-32ab-4b62-8d99-da8d5cfb7ad8","Type":"ContainerStarted","Data":"b24475bf0903ca124be07388898ceabcaa197269b184a281c81a5daaaf3d70ef"} Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.957016 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.960114 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" event={"ID":"692c8d95-4a19-48c8-b70d-571ffbb1ca8e","Type":"ContainerStarted","Data":"1982b7ac594995ed157faa0cbd756b7ae7c28ba52f4be942ab98c8698fa2695e"} Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.960182 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" event={"ID":"692c8d95-4a19-48c8-b70d-571ffbb1ca8e","Type":"ContainerStarted","Data":"d4a2cb14b630c139ec8b62781149735e04ea0beab94a09a69bcf87001c9e4a60"} Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.960504 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.961643 4954 patch_prober.go:28] interesting pod/controller-manager-868f679974-cxc7g container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.961691 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" podUID="692c8d95-4a19-48c8-b70d-571ffbb1ca8e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.967771 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.980116 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-544bb94698-rhrxj" podStartSLOduration=2.9800880530000002 podStartE2EDuration="2.980088053s" podCreationTimestamp="2025-12-09 17:02:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:02:48.973345226 +0000 UTC m=+365.361519056" watchObservedRunningTime="2025-12-09 17:02:48.980088053 +0000 UTC m=+365.368261883" Dec 09 17:02:48 crc kubenswrapper[4954]: I1209 17:02:48.994363 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" podStartSLOduration=2.99434039 podStartE2EDuration="2.99434039s" podCreationTimestamp="2025-12-09 17:02:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:02:48.991204599 +0000 UTC m=+365.379378459" watchObservedRunningTime="2025-12-09 17:02:48.99434039 +0000 UTC m=+365.382514230" Dec 09 17:02:49 crc kubenswrapper[4954]: I1209 17:02:49.971270 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:03:06 crc kubenswrapper[4954]: I1209 17:03:06.555111 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bg5hn" Dec 09 17:03:06 crc kubenswrapper[4954]: I1209 17:03:06.628509 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dx5bf"] Dec 09 17:03:13 crc kubenswrapper[4954]: I1209 17:03:13.755407 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:03:13 crc kubenswrapper[4954]: I1209 17:03:13.756500 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:03:26 crc kubenswrapper[4954]: I1209 17:03:26.923934 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-868f679974-cxc7g"] Dec 09 17:03:26 crc kubenswrapper[4954]: I1209 17:03:26.924748 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" podUID="692c8d95-4a19-48c8-b70d-571ffbb1ca8e" containerName="controller-manager" containerID="cri-o://1982b7ac594995ed157faa0cbd756b7ae7c28ba52f4be942ab98c8698fa2695e" gracePeriod=30 Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.196994 4954 generic.go:334] "Generic (PLEG): container finished" podID="692c8d95-4a19-48c8-b70d-571ffbb1ca8e" containerID="1982b7ac594995ed157faa0cbd756b7ae7c28ba52f4be942ab98c8698fa2695e" exitCode=0 Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.197040 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" event={"ID":"692c8d95-4a19-48c8-b70d-571ffbb1ca8e","Type":"ContainerDied","Data":"1982b7ac594995ed157faa0cbd756b7ae7c28ba52f4be942ab98c8698fa2695e"} Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.281959 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.446020 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-serving-cert\") pod \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.446239 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-config\") pod \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.446292 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-client-ca\") pod \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.446381 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-proxy-ca-bundles\") pod \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.446495 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hjvc\" (UniqueName: \"kubernetes.io/projected/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-kube-api-access-5hjvc\") pod \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\" (UID: \"692c8d95-4a19-48c8-b70d-571ffbb1ca8e\") " Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.447484 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-client-ca" (OuterVolumeSpecName: "client-ca") pod "692c8d95-4a19-48c8-b70d-571ffbb1ca8e" (UID: "692c8d95-4a19-48c8-b70d-571ffbb1ca8e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.447633 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-config" (OuterVolumeSpecName: "config") pod "692c8d95-4a19-48c8-b70d-571ffbb1ca8e" (UID: "692c8d95-4a19-48c8-b70d-571ffbb1ca8e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.447813 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "692c8d95-4a19-48c8-b70d-571ffbb1ca8e" (UID: "692c8d95-4a19-48c8-b70d-571ffbb1ca8e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.453124 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-kube-api-access-5hjvc" (OuterVolumeSpecName: "kube-api-access-5hjvc") pod "692c8d95-4a19-48c8-b70d-571ffbb1ca8e" (UID: "692c8d95-4a19-48c8-b70d-571ffbb1ca8e"). InnerVolumeSpecName "kube-api-access-5hjvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.453279 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "692c8d95-4a19-48c8-b70d-571ffbb1ca8e" (UID: "692c8d95-4a19-48c8-b70d-571ffbb1ca8e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.548675 4954 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.548719 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.548732 4954 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.548749 4954 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:27 crc kubenswrapper[4954]: I1209 17:03:27.548765 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hjvc\" (UniqueName: \"kubernetes.io/projected/692c8d95-4a19-48c8-b70d-571ffbb1ca8e-kube-api-access-5hjvc\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.204050 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" event={"ID":"692c8d95-4a19-48c8-b70d-571ffbb1ca8e","Type":"ContainerDied","Data":"d4a2cb14b630c139ec8b62781149735e04ea0beab94a09a69bcf87001c9e4a60"} Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.204163 4954 scope.go:117] "RemoveContainer" containerID="1982b7ac594995ed157faa0cbd756b7ae7c28ba52f4be942ab98c8698fa2695e" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.204319 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-868f679974-cxc7g" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.230442 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-868f679974-cxc7g"] Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.234117 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-868f679974-cxc7g"] Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.321985 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-77775444d6-lwt2m"] Dec 09 17:03:28 crc kubenswrapper[4954]: E1209 17:03:28.322249 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="692c8d95-4a19-48c8-b70d-571ffbb1ca8e" containerName="controller-manager" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.322263 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="692c8d95-4a19-48c8-b70d-571ffbb1ca8e" containerName="controller-manager" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.322390 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="692c8d95-4a19-48c8-b70d-571ffbb1ca8e" containerName="controller-manager" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.323062 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.326248 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.326495 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.326565 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.328844 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.331152 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.331351 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.335539 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77775444d6-lwt2m"] Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.336485 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.363272 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9083fc65-41ec-4cf1-be52-7749314221d2-proxy-ca-bundles\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.363334 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9083fc65-41ec-4cf1-be52-7749314221d2-serving-cert\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.363395 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9083fc65-41ec-4cf1-be52-7749314221d2-client-ca\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.363419 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gw62\" (UniqueName: \"kubernetes.io/projected/9083fc65-41ec-4cf1-be52-7749314221d2-kube-api-access-6gw62\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.363439 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9083fc65-41ec-4cf1-be52-7749314221d2-config\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.464051 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9083fc65-41ec-4cf1-be52-7749314221d2-proxy-ca-bundles\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.464107 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9083fc65-41ec-4cf1-be52-7749314221d2-serving-cert\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.464161 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9083fc65-41ec-4cf1-be52-7749314221d2-client-ca\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.464184 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gw62\" (UniqueName: \"kubernetes.io/projected/9083fc65-41ec-4cf1-be52-7749314221d2-kube-api-access-6gw62\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.464204 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9083fc65-41ec-4cf1-be52-7749314221d2-config\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.465888 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9083fc65-41ec-4cf1-be52-7749314221d2-config\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.465895 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9083fc65-41ec-4cf1-be52-7749314221d2-client-ca\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.466053 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9083fc65-41ec-4cf1-be52-7749314221d2-proxy-ca-bundles\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.468678 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9083fc65-41ec-4cf1-be52-7749314221d2-serving-cert\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.483890 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gw62\" (UniqueName: \"kubernetes.io/projected/9083fc65-41ec-4cf1-be52-7749314221d2-kube-api-access-6gw62\") pod \"controller-manager-77775444d6-lwt2m\" (UID: \"9083fc65-41ec-4cf1-be52-7749314221d2\") " pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.645820 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:28 crc kubenswrapper[4954]: I1209 17:03:28.835286 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77775444d6-lwt2m"] Dec 09 17:03:29 crc kubenswrapper[4954]: I1209 17:03:29.210618 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" event={"ID":"9083fc65-41ec-4cf1-be52-7749314221d2","Type":"ContainerStarted","Data":"34b17e4a66818d5b0aa04b6d7375644e6b2dd2e451a1248e452f513f3cea5b05"} Dec 09 17:03:29 crc kubenswrapper[4954]: I1209 17:03:29.210917 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" event={"ID":"9083fc65-41ec-4cf1-be52-7749314221d2","Type":"ContainerStarted","Data":"b7b88cb519d48ecd54fd44d3767a74b5d21dfb2824fdbbcb2ca3608f80d8b69f"} Dec 09 17:03:29 crc kubenswrapper[4954]: I1209 17:03:29.212088 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:29 crc kubenswrapper[4954]: I1209 17:03:29.236321 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" Dec 09 17:03:29 crc kubenswrapper[4954]: I1209 17:03:29.246398 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-77775444d6-lwt2m" podStartSLOduration=3.246374254 podStartE2EDuration="3.246374254s" podCreationTimestamp="2025-12-09 17:03:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:03:29.241944914 +0000 UTC m=+405.630118734" watchObservedRunningTime="2025-12-09 17:03:29.246374254 +0000 UTC m=+405.634548074" Dec 09 17:03:30 crc kubenswrapper[4954]: I1209 17:03:30.127121 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="692c8d95-4a19-48c8-b70d-571ffbb1ca8e" path="/var/lib/kubelet/pods/692c8d95-4a19-48c8-b70d-571ffbb1ca8e/volumes" Dec 09 17:03:31 crc kubenswrapper[4954]: I1209 17:03:31.669135 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" podUID="d284ca9d-999e-4ab5-8d7e-647af799d69c" containerName="registry" containerID="cri-o://dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952" gracePeriod=30 Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.109103 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.234121 4954 generic.go:334] "Generic (PLEG): container finished" podID="d284ca9d-999e-4ab5-8d7e-647af799d69c" containerID="dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952" exitCode=0 Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.234151 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" event={"ID":"d284ca9d-999e-4ab5-8d7e-647af799d69c","Type":"ContainerDied","Data":"dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952"} Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.234183 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" event={"ID":"d284ca9d-999e-4ab5-8d7e-647af799d69c","Type":"ContainerDied","Data":"04e8fd59b0bd3bb85eea438d9c0e8b48560d6ba53e15e3d073801b4529d045bb"} Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.234210 4954 scope.go:117] "RemoveContainer" containerID="dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.234254 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dx5bf" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.250676 4954 scope.go:117] "RemoveContainer" containerID="dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952" Dec 09 17:03:32 crc kubenswrapper[4954]: E1209 17:03:32.251203 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952\": container with ID starting with dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952 not found: ID does not exist" containerID="dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.251334 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952"} err="failed to get container status \"dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952\": rpc error: code = NotFound desc = could not find container \"dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952\": container with ID starting with dff0fa31af02a51a07d5045ba455e422b8282df1a4184ed9ef35315d33559952 not found: ID does not exist" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.268389 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-trusted-ca\") pod \"d284ca9d-999e-4ab5-8d7e-647af799d69c\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.268461 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-bound-sa-token\") pod \"d284ca9d-999e-4ab5-8d7e-647af799d69c\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.268507 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d284ca9d-999e-4ab5-8d7e-647af799d69c-installation-pull-secrets\") pod \"d284ca9d-999e-4ab5-8d7e-647af799d69c\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.268548 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-tls\") pod \"d284ca9d-999e-4ab5-8d7e-647af799d69c\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.268616 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d284ca9d-999e-4ab5-8d7e-647af799d69c-ca-trust-extracted\") pod \"d284ca9d-999e-4ab5-8d7e-647af799d69c\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.268657 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-certificates\") pod \"d284ca9d-999e-4ab5-8d7e-647af799d69c\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.268702 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n8gl\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-kube-api-access-9n8gl\") pod \"d284ca9d-999e-4ab5-8d7e-647af799d69c\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.269057 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d284ca9d-999e-4ab5-8d7e-647af799d69c\" (UID: \"d284ca9d-999e-4ab5-8d7e-647af799d69c\") " Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.269420 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d284ca9d-999e-4ab5-8d7e-647af799d69c" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.269505 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d284ca9d-999e-4ab5-8d7e-647af799d69c" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.269553 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.276903 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d284ca9d-999e-4ab5-8d7e-647af799d69c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d284ca9d-999e-4ab5-8d7e-647af799d69c" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.279072 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-kube-api-access-9n8gl" (OuterVolumeSpecName: "kube-api-access-9n8gl") pod "d284ca9d-999e-4ab5-8d7e-647af799d69c" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c"). InnerVolumeSpecName "kube-api-access-9n8gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.279427 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d284ca9d-999e-4ab5-8d7e-647af799d69c" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.279774 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d284ca9d-999e-4ab5-8d7e-647af799d69c" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.286270 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d284ca9d-999e-4ab5-8d7e-647af799d69c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d284ca9d-999e-4ab5-8d7e-647af799d69c" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.288635 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d284ca9d-999e-4ab5-8d7e-647af799d69c" (UID: "d284ca9d-999e-4ab5-8d7e-647af799d69c"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.371164 4954 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.371529 4954 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d284ca9d-999e-4ab5-8d7e-647af799d69c-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.371550 4954 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.371560 4954 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d284ca9d-999e-4ab5-8d7e-647af799d69c-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.371571 4954 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d284ca9d-999e-4ab5-8d7e-647af799d69c-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.371580 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n8gl\" (UniqueName: \"kubernetes.io/projected/d284ca9d-999e-4ab5-8d7e-647af799d69c-kube-api-access-9n8gl\") on node \"crc\" DevicePath \"\"" Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.568621 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dx5bf"] Dec 09 17:03:32 crc kubenswrapper[4954]: I1209 17:03:32.573711 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dx5bf"] Dec 09 17:03:34 crc kubenswrapper[4954]: I1209 17:03:34.135422 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d284ca9d-999e-4ab5-8d7e-647af799d69c" path="/var/lib/kubelet/pods/d284ca9d-999e-4ab5-8d7e-647af799d69c/volumes" Dec 09 17:03:43 crc kubenswrapper[4954]: I1209 17:03:43.755573 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:03:43 crc kubenswrapper[4954]: I1209 17:03:43.756105 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:03:43 crc kubenswrapper[4954]: I1209 17:03:43.756143 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:03:43 crc kubenswrapper[4954]: I1209 17:03:43.756716 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33a69fad0fc5c3316079e821a4a615a1b8c30db3b75d7d52d707b8d7b04f4a32"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:03:43 crc kubenswrapper[4954]: I1209 17:03:43.756786 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://33a69fad0fc5c3316079e821a4a615a1b8c30db3b75d7d52d707b8d7b04f4a32" gracePeriod=600 Dec 09 17:03:44 crc kubenswrapper[4954]: I1209 17:03:44.302081 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="33a69fad0fc5c3316079e821a4a615a1b8c30db3b75d7d52d707b8d7b04f4a32" exitCode=0 Dec 09 17:03:44 crc kubenswrapper[4954]: I1209 17:03:44.302156 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"33a69fad0fc5c3316079e821a4a615a1b8c30db3b75d7d52d707b8d7b04f4a32"} Dec 09 17:03:44 crc kubenswrapper[4954]: I1209 17:03:44.303106 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"f40e602b695fb058005b96319e15b3099ce90e1eb551fb5c6fef94df1b5d64c9"} Dec 09 17:03:44 crc kubenswrapper[4954]: I1209 17:03:44.303298 4954 scope.go:117] "RemoveContainer" containerID="c1e1b734f65f85e477d84baf3c5dec74a9f3c2079f39a77d0b854b1e586e6dc2" Dec 09 17:06:13 crc kubenswrapper[4954]: I1209 17:06:13.755341 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:06:13 crc kubenswrapper[4954]: I1209 17:06:13.756376 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:06:43 crc kubenswrapper[4954]: I1209 17:06:43.755323 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:06:43 crc kubenswrapper[4954]: I1209 17:06:43.755871 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:07:13 crc kubenswrapper[4954]: I1209 17:07:13.755277 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:07:13 crc kubenswrapper[4954]: I1209 17:07:13.755812 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:07:13 crc kubenswrapper[4954]: I1209 17:07:13.755860 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:07:13 crc kubenswrapper[4954]: I1209 17:07:13.756474 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f40e602b695fb058005b96319e15b3099ce90e1eb551fb5c6fef94df1b5d64c9"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:07:13 crc kubenswrapper[4954]: I1209 17:07:13.756521 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://f40e602b695fb058005b96319e15b3099ce90e1eb551fb5c6fef94df1b5d64c9" gracePeriod=600 Dec 09 17:07:14 crc kubenswrapper[4954]: I1209 17:07:14.626741 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="f40e602b695fb058005b96319e15b3099ce90e1eb551fb5c6fef94df1b5d64c9" exitCode=0 Dec 09 17:07:14 crc kubenswrapper[4954]: I1209 17:07:14.626803 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"f40e602b695fb058005b96319e15b3099ce90e1eb551fb5c6fef94df1b5d64c9"} Dec 09 17:07:14 crc kubenswrapper[4954]: I1209 17:07:14.627093 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"ea6d5b701a579f932d38d9e389879e05d1afbb8ef2ed598058ec1bc38f02d1f5"} Dec 09 17:07:14 crc kubenswrapper[4954]: I1209 17:07:14.627115 4954 scope.go:117] "RemoveContainer" containerID="33a69fad0fc5c3316079e821a4a615a1b8c30db3b75d7d52d707b8d7b04f4a32" Dec 09 17:07:34 crc kubenswrapper[4954]: I1209 17:07:34.976129 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5"] Dec 09 17:07:34 crc kubenswrapper[4954]: E1209 17:07:34.977491 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d284ca9d-999e-4ab5-8d7e-647af799d69c" containerName="registry" Dec 09 17:07:34 crc kubenswrapper[4954]: I1209 17:07:34.977514 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d284ca9d-999e-4ab5-8d7e-647af799d69c" containerName="registry" Dec 09 17:07:34 crc kubenswrapper[4954]: I1209 17:07:34.977727 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d284ca9d-999e-4ab5-8d7e-647af799d69c" containerName="registry" Dec 09 17:07:34 crc kubenswrapper[4954]: I1209 17:07:34.979078 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:34 crc kubenswrapper[4954]: I1209 17:07:34.982745 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 17:07:34 crc kubenswrapper[4954]: I1209 17:07:34.985713 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5"] Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.122639 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.122709 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbzmh\" (UniqueName: \"kubernetes.io/projected/39ee90e9-4a6f-4f98-9dab-da37a93221f2-kube-api-access-mbzmh\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.122752 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.223982 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.224040 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbzmh\" (UniqueName: \"kubernetes.io/projected/39ee90e9-4a6f-4f98-9dab-da37a93221f2-kube-api-access-mbzmh\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.224084 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.224576 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.224692 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.244624 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbzmh\" (UniqueName: \"kubernetes.io/projected/39ee90e9-4a6f-4f98-9dab-da37a93221f2-kube-api-access-mbzmh\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.305365 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.490528 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5"] Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.749195 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" event={"ID":"39ee90e9-4a6f-4f98-9dab-da37a93221f2","Type":"ContainerStarted","Data":"dc87a87a32181d25699fbfe57266b887400734344545a9da5b532a99eedaca0c"} Dec 09 17:07:35 crc kubenswrapper[4954]: I1209 17:07:35.749274 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" event={"ID":"39ee90e9-4a6f-4f98-9dab-da37a93221f2","Type":"ContainerStarted","Data":"6e744c8b59c1b586bd13ba2383106e4a1e7c3f88f944f8e7ca25418a42416412"} Dec 09 17:07:36 crc kubenswrapper[4954]: I1209 17:07:36.758103 4954 generic.go:334] "Generic (PLEG): container finished" podID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" containerID="dc87a87a32181d25699fbfe57266b887400734344545a9da5b532a99eedaca0c" exitCode=0 Dec 09 17:07:36 crc kubenswrapper[4954]: I1209 17:07:36.758150 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" event={"ID":"39ee90e9-4a6f-4f98-9dab-da37a93221f2","Type":"ContainerDied","Data":"dc87a87a32181d25699fbfe57266b887400734344545a9da5b532a99eedaca0c"} Dec 09 17:07:36 crc kubenswrapper[4954]: I1209 17:07:36.761025 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 17:07:38 crc kubenswrapper[4954]: I1209 17:07:38.771236 4954 generic.go:334] "Generic (PLEG): container finished" podID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" containerID="687ca25af671c94a5c36d52a18645371434b84712dda6eefb6d80d456df6e1f3" exitCode=0 Dec 09 17:07:38 crc kubenswrapper[4954]: I1209 17:07:38.771334 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" event={"ID":"39ee90e9-4a6f-4f98-9dab-da37a93221f2","Type":"ContainerDied","Data":"687ca25af671c94a5c36d52a18645371434b84712dda6eefb6d80d456df6e1f3"} Dec 09 17:07:39 crc kubenswrapper[4954]: I1209 17:07:39.783962 4954 generic.go:334] "Generic (PLEG): container finished" podID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" containerID="a5d1719cf6f2890787807b62cb8048e90a86df344d1a2e7f1e476fbea320c809" exitCode=0 Dec 09 17:07:39 crc kubenswrapper[4954]: I1209 17:07:39.784088 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" event={"ID":"39ee90e9-4a6f-4f98-9dab-da37a93221f2","Type":"ContainerDied","Data":"a5d1719cf6f2890787807b62cb8048e90a86df344d1a2e7f1e476fbea320c809"} Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.025931 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.125280 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbzmh\" (UniqueName: \"kubernetes.io/projected/39ee90e9-4a6f-4f98-9dab-da37a93221f2-kube-api-access-mbzmh\") pod \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.125564 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-bundle\") pod \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.127468 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-bundle" (OuterVolumeSpecName: "bundle") pod "39ee90e9-4a6f-4f98-9dab-da37a93221f2" (UID: "39ee90e9-4a6f-4f98-9dab-da37a93221f2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.131268 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39ee90e9-4a6f-4f98-9dab-da37a93221f2-kube-api-access-mbzmh" (OuterVolumeSpecName: "kube-api-access-mbzmh") pod "39ee90e9-4a6f-4f98-9dab-da37a93221f2" (UID: "39ee90e9-4a6f-4f98-9dab-da37a93221f2"). InnerVolumeSpecName "kube-api-access-mbzmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.226470 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-util\") pod \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\" (UID: \"39ee90e9-4a6f-4f98-9dab-da37a93221f2\") " Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.226917 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbzmh\" (UniqueName: \"kubernetes.io/projected/39ee90e9-4a6f-4f98-9dab-da37a93221f2-kube-api-access-mbzmh\") on node \"crc\" DevicePath \"\"" Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.226946 4954 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.438073 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-util" (OuterVolumeSpecName: "util") pod "39ee90e9-4a6f-4f98-9dab-da37a93221f2" (UID: "39ee90e9-4a6f-4f98-9dab-da37a93221f2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.530624 4954 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/39ee90e9-4a6f-4f98-9dab-da37a93221f2-util\") on node \"crc\" DevicePath \"\"" Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.799376 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" event={"ID":"39ee90e9-4a6f-4f98-9dab-da37a93221f2","Type":"ContainerDied","Data":"6e744c8b59c1b586bd13ba2383106e4a1e7c3f88f944f8e7ca25418a42416412"} Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.799432 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e744c8b59c1b586bd13ba2383106e4a1e7c3f88f944f8e7ca25418a42416412" Dec 09 17:07:41 crc kubenswrapper[4954]: I1209 17:07:41.799444 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.107783 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn"] Dec 09 17:07:55 crc kubenswrapper[4954]: E1209 17:07:55.109547 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" containerName="pull" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.109572 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" containerName="pull" Dec 09 17:07:55 crc kubenswrapper[4954]: E1209 17:07:55.109616 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" containerName="extract" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.109626 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" containerName="extract" Dec 09 17:07:55 crc kubenswrapper[4954]: E1209 17:07:55.109652 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" containerName="util" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.109660 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" containerName="util" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.109791 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" containerName="extract" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.110417 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.119616 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn"] Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.119963 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.120212 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-lzcvf" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.120471 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.194939 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q8nm\" (UniqueName: \"kubernetes.io/projected/0e70cc7a-1b64-47c6-b0f1-4da0e77c9870-kube-api-access-4q8nm\") pod \"obo-prometheus-operator-668cf9dfbb-gtwjn\" (UID: \"0e70cc7a-1b64-47c6-b0f1-4da0e77c9870\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.298449 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q8nm\" (UniqueName: \"kubernetes.io/projected/0e70cc7a-1b64-47c6-b0f1-4da0e77c9870-kube-api-access-4q8nm\") pod \"obo-prometheus-operator-668cf9dfbb-gtwjn\" (UID: \"0e70cc7a-1b64-47c6-b0f1-4da0e77c9870\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.330290 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q8nm\" (UniqueName: \"kubernetes.io/projected/0e70cc7a-1b64-47c6-b0f1-4da0e77c9870-kube-api-access-4q8nm\") pod \"obo-prometheus-operator-668cf9dfbb-gtwjn\" (UID: \"0e70cc7a-1b64-47c6-b0f1-4da0e77c9870\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.335535 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn"] Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.336424 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.339086 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-cvtdb" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.339331 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.345877 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m"] Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.346634 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.400258 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/71312e1a-0fc4-417a-92e3-080c5651ff47-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn\" (UID: \"71312e1a-0fc4-417a-92e3-080c5651ff47\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.400638 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/71312e1a-0fc4-417a-92e3-080c5651ff47-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn\" (UID: \"71312e1a-0fc4-417a-92e3-080c5651ff47\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.400666 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab2760a1-3695-4556-bff0-d61d9da9d78b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m\" (UID: \"ab2760a1-3695-4556-bff0-d61d9da9d78b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.400690 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab2760a1-3695-4556-bff0-d61d9da9d78b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m\" (UID: \"ab2760a1-3695-4556-bff0-d61d9da9d78b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.444456 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m"] Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.450983 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn"] Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.503622 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/71312e1a-0fc4-417a-92e3-080c5651ff47-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn\" (UID: \"71312e1a-0fc4-417a-92e3-080c5651ff47\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.503733 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab2760a1-3695-4556-bff0-d61d9da9d78b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m\" (UID: \"ab2760a1-3695-4556-bff0-d61d9da9d78b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.503770 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab2760a1-3695-4556-bff0-d61d9da9d78b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m\" (UID: \"ab2760a1-3695-4556-bff0-d61d9da9d78b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.503849 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/71312e1a-0fc4-417a-92e3-080c5651ff47-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn\" (UID: \"71312e1a-0fc4-417a-92e3-080c5651ff47\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.508011 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-87ttl"] Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.508331 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/71312e1a-0fc4-417a-92e3-080c5651ff47-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn\" (UID: \"71312e1a-0fc4-417a-92e3-080c5651ff47\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.511283 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab2760a1-3695-4556-bff0-d61d9da9d78b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m\" (UID: \"ab2760a1-3695-4556-bff0-d61d9da9d78b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.511729 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.513958 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/71312e1a-0fc4-417a-92e3-080c5651ff47-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn\" (UID: \"71312e1a-0fc4-417a-92e3-080c5651ff47\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.515411 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-s2qkb" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.515799 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.517410 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.526402 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab2760a1-3695-4556-bff0-d61d9da9d78b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m\" (UID: \"ab2760a1-3695-4556-bff0-d61d9da9d78b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.530675 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-87ttl"] Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.605126 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-87ttl\" (UID: \"cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6\") " pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.605232 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m42l6\" (UniqueName: \"kubernetes.io/projected/cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6-kube-api-access-m42l6\") pod \"observability-operator-d8bb48f5d-87ttl\" (UID: \"cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6\") " pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.667342 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.677021 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.682842 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-2l8fr"] Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.683513 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-2l8fr" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.688580 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-wb6p7" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.706810 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq44b\" (UniqueName: \"kubernetes.io/projected/2466d04c-561e-4c13-9939-9194d26ac50f-kube-api-access-sq44b\") pod \"perses-operator-5446b9c989-2l8fr\" (UID: \"2466d04c-561e-4c13-9939-9194d26ac50f\") " pod="openshift-operators/perses-operator-5446b9c989-2l8fr" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.706865 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m42l6\" (UniqueName: \"kubernetes.io/projected/cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6-kube-api-access-m42l6\") pod \"observability-operator-d8bb48f5d-87ttl\" (UID: \"cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6\") " pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.706894 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2466d04c-561e-4c13-9939-9194d26ac50f-openshift-service-ca\") pod \"perses-operator-5446b9c989-2l8fr\" (UID: \"2466d04c-561e-4c13-9939-9194d26ac50f\") " pod="openshift-operators/perses-operator-5446b9c989-2l8fr" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.706961 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-87ttl\" (UID: \"cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6\") " pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.708034 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-2l8fr"] Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.721152 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-87ttl\" (UID: \"cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6\") " pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.771093 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m42l6\" (UniqueName: \"kubernetes.io/projected/cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6-kube-api-access-m42l6\") pod \"observability-operator-d8bb48f5d-87ttl\" (UID: \"cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6\") " pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.811003 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq44b\" (UniqueName: \"kubernetes.io/projected/2466d04c-561e-4c13-9939-9194d26ac50f-kube-api-access-sq44b\") pod \"perses-operator-5446b9c989-2l8fr\" (UID: \"2466d04c-561e-4c13-9939-9194d26ac50f\") " pod="openshift-operators/perses-operator-5446b9c989-2l8fr" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.811075 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2466d04c-561e-4c13-9939-9194d26ac50f-openshift-service-ca\") pod \"perses-operator-5446b9c989-2l8fr\" (UID: \"2466d04c-561e-4c13-9939-9194d26ac50f\") " pod="openshift-operators/perses-operator-5446b9c989-2l8fr" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.812121 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/2466d04c-561e-4c13-9939-9194d26ac50f-openshift-service-ca\") pod \"perses-operator-5446b9c989-2l8fr\" (UID: \"2466d04c-561e-4c13-9939-9194d26ac50f\") " pod="openshift-operators/perses-operator-5446b9c989-2l8fr" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.849099 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq44b\" (UniqueName: \"kubernetes.io/projected/2466d04c-561e-4c13-9939-9194d26ac50f-kube-api-access-sq44b\") pod \"perses-operator-5446b9c989-2l8fr\" (UID: \"2466d04c-561e-4c13-9939-9194d26ac50f\") " pod="openshift-operators/perses-operator-5446b9c989-2l8fr" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.916170 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" Dec 09 17:07:55 crc kubenswrapper[4954]: I1209 17:07:55.932810 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn"] Dec 09 17:07:56 crc kubenswrapper[4954]: I1209 17:07:56.023521 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn"] Dec 09 17:07:56 crc kubenswrapper[4954]: I1209 17:07:56.058964 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn" event={"ID":"0e70cc7a-1b64-47c6-b0f1-4da0e77c9870","Type":"ContainerStarted","Data":"f109fed568636f2186681f0692cf5a02f813c91c4b8932829b4ea6c5e67bd1e0"} Dec 09 17:07:56 crc kubenswrapper[4954]: I1209 17:07:56.060555 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" event={"ID":"71312e1a-0fc4-417a-92e3-080c5651ff47","Type":"ContainerStarted","Data":"f4d2a922868129a369c77a0ca5f4d927f16c371aebf6bb95208004603e1b2bd7"} Dec 09 17:07:56 crc kubenswrapper[4954]: I1209 17:07:56.070034 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-2l8fr" Dec 09 17:07:56 crc kubenswrapper[4954]: I1209 17:07:56.071137 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m"] Dec 09 17:07:56 crc kubenswrapper[4954]: W1209 17:07:56.095368 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab2760a1_3695_4556_bff0_d61d9da9d78b.slice/crio-858ec8b1dcd10999ae6fa52ec6ee69d8b2c0336c764ae0e3527f23bb40599d03 WatchSource:0}: Error finding container 858ec8b1dcd10999ae6fa52ec6ee69d8b2c0336c764ae0e3527f23bb40599d03: Status 404 returned error can't find the container with id 858ec8b1dcd10999ae6fa52ec6ee69d8b2c0336c764ae0e3527f23bb40599d03 Dec 09 17:07:56 crc kubenswrapper[4954]: I1209 17:07:56.240434 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-87ttl"] Dec 09 17:07:56 crc kubenswrapper[4954]: I1209 17:07:56.434546 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-2l8fr"] Dec 09 17:07:57 crc kubenswrapper[4954]: I1209 17:07:57.068909 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-2l8fr" event={"ID":"2466d04c-561e-4c13-9939-9194d26ac50f","Type":"ContainerStarted","Data":"3cc94b09f80be8b4481faf34a615f9be60c86ab687b66830d503035d18c3deda"} Dec 09 17:07:57 crc kubenswrapper[4954]: I1209 17:07:57.070630 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" event={"ID":"ab2760a1-3695-4556-bff0-d61d9da9d78b","Type":"ContainerStarted","Data":"858ec8b1dcd10999ae6fa52ec6ee69d8b2c0336c764ae0e3527f23bb40599d03"} Dec 09 17:07:57 crc kubenswrapper[4954]: I1209 17:07:57.071973 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" event={"ID":"cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6","Type":"ContainerStarted","Data":"c36656027a362c098086537e33c65a355462241505a59121fc6832c1663a9977"} Dec 09 17:08:04 crc kubenswrapper[4954]: I1209 17:08:04.307655 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7b4bw"] Dec 09 17:08:04 crc kubenswrapper[4954]: I1209 17:08:04.308997 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovn-controller" containerID="cri-o://503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6" gracePeriod=30 Dec 09 17:08:04 crc kubenswrapper[4954]: I1209 17:08:04.309353 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="northd" containerID="cri-o://194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf" gracePeriod=30 Dec 09 17:08:04 crc kubenswrapper[4954]: I1209 17:08:04.309753 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e" gracePeriod=30 Dec 09 17:08:04 crc kubenswrapper[4954]: I1209 17:08:04.310008 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovn-acl-logging" containerID="cri-o://933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e" gracePeriod=30 Dec 09 17:08:04 crc kubenswrapper[4954]: I1209 17:08:04.310216 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="kube-rbac-proxy-node" containerID="cri-o://5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb" gracePeriod=30 Dec 09 17:08:04 crc kubenswrapper[4954]: I1209 17:08:04.310308 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="sbdb" containerID="cri-o://83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807" gracePeriod=30 Dec 09 17:08:04 crc kubenswrapper[4954]: I1209 17:08:04.310385 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="nbdb" containerID="cri-o://786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845" gracePeriod=30 Dec 09 17:08:04 crc kubenswrapper[4954]: I1209 17:08:04.393092 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" containerID="cri-o://03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc" gracePeriod=30 Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.179358 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9n2h_1eccf2e0-30a3-4201-a23f-85a92a522d72/kube-multus/2.log" Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.180976 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9n2h_1eccf2e0-30a3-4201-a23f-85a92a522d72/kube-multus/1.log" Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.181019 4954 generic.go:334] "Generic (PLEG): container finished" podID="1eccf2e0-30a3-4201-a23f-85a92a522d72" containerID="e0c232380ddfbb380b2f38c01d48754a185f0b2cee838436385d29e2945680e0" exitCode=2 Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.181097 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9n2h" event={"ID":"1eccf2e0-30a3-4201-a23f-85a92a522d72","Type":"ContainerDied","Data":"e0c232380ddfbb380b2f38c01d48754a185f0b2cee838436385d29e2945680e0"} Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.181143 4954 scope.go:117] "RemoveContainer" containerID="adbd985ee682b5c18a20890f8ba14970203478e3c567ff663cae5b3cd8392010" Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.181581 4954 scope.go:117] "RemoveContainer" containerID="e0c232380ddfbb380b2f38c01d48754a185f0b2cee838436385d29e2945680e0" Dec 09 17:08:05 crc kubenswrapper[4954]: E1209 17:08:05.181865 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-h9n2h_openshift-multus(1eccf2e0-30a3-4201-a23f-85a92a522d72)\"" pod="openshift-multus/multus-h9n2h" podUID="1eccf2e0-30a3-4201-a23f-85a92a522d72" Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.187041 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovnkube-controller/3.log" Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.197275 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovn-acl-logging/0.log" Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.197887 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovn-controller/0.log" Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198654 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc" exitCode=0 Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198677 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807" exitCode=0 Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198699 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845" exitCode=0 Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198705 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf" exitCode=0 Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198711 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e" exitCode=0 Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198717 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb" exitCode=0 Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198724 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e" exitCode=143 Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198730 4954 generic.go:334] "Generic (PLEG): container finished" podID="5736742d-6a6c-4309-b95d-6103408901ec" containerID="503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6" exitCode=143 Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198751 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc"} Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198793 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807"} Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198803 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845"} Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198811 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf"} Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198819 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e"} Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198828 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb"} Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198937 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e"} Dec 09 17:08:05 crc kubenswrapper[4954]: I1209 17:08:05.198952 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6"} Dec 09 17:08:12 crc kubenswrapper[4954]: E1209 17:08:12.241907 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc is running failed: container process not found" containerID="03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Dec 09 17:08:12 crc kubenswrapper[4954]: E1209 17:08:12.243094 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc is running failed: container process not found" containerID="03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Dec 09 17:08:12 crc kubenswrapper[4954]: E1209 17:08:12.243452 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc is running failed: container process not found" containerID="03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Dec 09 17:08:12 crc kubenswrapper[4954]: E1209 17:08:12.243533 4954 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:13 crc kubenswrapper[4954]: E1209 17:08:13.444901 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845 is running failed: container process not found" containerID="786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 09 17:08:13 crc kubenswrapper[4954]: E1209 17:08:13.445078 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807 is running failed: container process not found" containerID="83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 09 17:08:13 crc kubenswrapper[4954]: E1209 17:08:13.445239 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845 is running failed: container process not found" containerID="786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 09 17:08:13 crc kubenswrapper[4954]: E1209 17:08:13.445329 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807 is running failed: container process not found" containerID="83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 09 17:08:13 crc kubenswrapper[4954]: E1209 17:08:13.445511 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845 is running failed: container process not found" containerID="786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 09 17:08:13 crc kubenswrapper[4954]: E1209 17:08:13.445939 4954 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="nbdb" Dec 09 17:08:13 crc kubenswrapper[4954]: E1209 17:08:13.446058 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807 is running failed: container process not found" containerID="83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 09 17:08:13 crc kubenswrapper[4954]: E1209 17:08:13.446088 4954 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="sbdb" Dec 09 17:08:15 crc kubenswrapper[4954]: E1209 17:08:15.803252 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb" Dec 09 17:08:15 crc kubenswrapper[4954]: E1209 17:08:15.803563 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb,Command:[],Args:[--namespace=$(NAMESPACE) --images=perses=$(RELATED_IMAGE_PERSES) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:e718854a7d6ca8accf0fa72db0eb902e46c44d747ad51dc3f06bba0cefaa3c01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:17ea20be390a94ab39f5cdd7f0cbc2498046eebcf77fe3dec9aa288d5c2cf46b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:d972f4faa5e9c121402d23ed85002f26af48ec36b1b71a7489d677b3913d08b4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-rhel9@sha256:91531137fc1dcd740e277e0f65e120a0176a16f788c14c27925b61aa0b792ade,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-rhel9@sha256:a69da8bbca8a28dd2925f864d51cc31cf761b10532c553095ba40b242ef701cb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-rhel9@sha256:897e1bfad1187062725b54d87107bd0155972257a50d8335dd29e1999b828a4f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf5-rhel9@sha256:95fe5b5746ca8c07ac9217ce2d8ac8e6afad17af210f9d8e0074df1310b209a8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf4-rhel9@sha256:e9d9a89e4d8126a62b1852055482258ee528cac6398dd5d43ebad75ace0f33c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-rhel9@sha256:ec684a0645ceb917b019af7ddba68c3533416e356ab0d0320a30e75ca7ebb31b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-pf4-rhel9@sha256:3b9693fcde9b3a9494fb04735b1f7cfd0426f10be820fdc3f024175c0d3df1c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-rhel9@sha256:580606f194180accc8abba099e17a26dca7522ec6d233fa2fdd40312771703e3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-rhel9@sha256:e03777be39e71701935059cd877603874a13ac94daa73219d4e5e545599d78a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-pf5-rhel9@sha256:aa47256193cfd2877853878e1ae97d2ab8b8e5deae62b387cbfad02b284d379c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:c595ff56b2cb85514bf4784db6ddb82e4e657e3e708a7fb695fc4997379a94d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:45a4ec2a519bcec99e886aa91596d5356a2414a2bd103baaef9fa7838c672eb2,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m42l6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-d8bb48f5d-87ttl_openshift-operators(cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:08:15 crc kubenswrapper[4954]: E1209 17:08:15.804817 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" podUID="cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6" Dec 09 17:08:16 crc kubenswrapper[4954]: E1209 17:08:16.267790 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb\\\"\"" pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" podUID="cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6" Dec 09 17:08:16 crc kubenswrapper[4954]: E1209 17:08:16.511641 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 09 17:08:16 crc kubenswrapper[4954]: E1209 17:08:16.511823 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m_openshift-operators(ab2760a1-3695-4556-bff0-d61d9da9d78b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:08:16 crc kubenswrapper[4954]: E1209 17:08:16.513071 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" podUID="ab2760a1-3695-4556-bff0-d61d9da9d78b" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.107050 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.107730 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn_openshift-operators(71312e1a-0fc4-417a-92e3-080c5651ff47): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.108995 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" podUID="71312e1a-0fc4-417a-92e3-080c5651ff47" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.113110 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.113279 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sq44b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5446b9c989-2l8fr_openshift-operators(2466d04c-561e-4c13-9939-9194d26ac50f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.114472 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5446b9c989-2l8fr" podUID="2466d04c-561e-4c13-9939-9194d26ac50f" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.122771 4954 scope.go:117] "RemoveContainer" containerID="e0c232380ddfbb380b2f38c01d48754a185f0b2cee838436385d29e2945680e0" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.122998 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-h9n2h_openshift-multus(1eccf2e0-30a3-4201-a23f-85a92a522d72)\"" pod="openshift-multus/multus-h9n2h" podUID="1eccf2e0-30a3-4201-a23f-85a92a522d72" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.144645 4954 scope.go:117] "RemoveContainer" containerID="6dccfa6101d40eaa4b3aaf917015829b0118418bf6a7023b2cebe8dd04c0f9b3" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.149111 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovn-acl-logging/0.log" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.149518 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovn-controller/0.log" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.149890 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.222809 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qjqcn"] Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223111 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="kube-rbac-proxy-node" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223129 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="kube-rbac-proxy-node" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223142 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223149 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223160 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223167 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223180 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="northd" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223187 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="northd" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223194 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223200 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223207 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="sbdb" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223212 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="sbdb" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223222 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="nbdb" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223227 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="nbdb" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223235 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="kubecfg-setup" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223241 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="kubecfg-setup" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223250 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223255 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223265 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovn-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223272 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovn-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223282 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovn-acl-logging" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223289 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovn-acl-logging" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223543 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223553 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="kube-rbac-proxy-node" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223561 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="nbdb" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223570 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223577 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223584 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="northd" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223603 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223608 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovn-acl-logging" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223615 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="sbdb" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223622 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovn-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223734 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223742 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.223750 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223756 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.223845 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.224013 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5736742d-6a6c-4309-b95d-6103408901ec" containerName="ovnkube-controller" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.226112 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.261862 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-netns\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.261902 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-ovn-kubernetes\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.261942 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-var-lib-cni-networks-ovn-kubernetes\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.261973 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5736742d-6a6c-4309-b95d-6103408901ec-ovn-node-metrics-cert\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.261995 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-netd\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262008 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-openvswitch\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262008 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262002 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262025 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-systemd\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262065 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262079 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262115 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msqm2\" (UniqueName: \"kubernetes.io/projected/5736742d-6a6c-4309-b95d-6103408901ec-kube-api-access-msqm2\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262124 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262147 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-bin\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262174 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-script-lib\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262190 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262221 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-ovn\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262257 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-env-overrides\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262275 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-node-log\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262284 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262302 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-systemd-units\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262326 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-kubelet\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262332 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-node-log" (OuterVolumeSpecName: "node-log") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262357 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-config\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262367 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262394 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262399 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-var-lib-openvswitch\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262436 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-slash\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262461 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-log-socket\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262519 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-etc-openvswitch\") pod \"5736742d-6a6c-4309-b95d-6103408901ec\" (UID: \"5736742d-6a6c-4309-b95d-6103408901ec\") " Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262795 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262813 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262835 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-log-socket" (OuterVolumeSpecName: "log-socket") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262848 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.262868 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263073 4954 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263089 4954 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263102 4954 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263118 4954 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263130 4954 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263140 4954 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263151 4954 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263160 4954 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263170 4954 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263180 4954 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-node-log\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263190 4954 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263200 4954 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263210 4954 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263222 4954 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263231 4954 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-log-socket\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263435 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-slash" (OuterVolumeSpecName: "host-slash") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.263785 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.267466 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5736742d-6a6c-4309-b95d-6103408901ec-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.267862 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5736742d-6a6c-4309-b95d-6103408901ec-kube-api-access-msqm2" (OuterVolumeSpecName: "kube-api-access-msqm2") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "kube-api-access-msqm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.277103 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9n2h_1eccf2e0-30a3-4201-a23f-85a92a522d72/kube-multus/2.log" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.283122 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "5736742d-6a6c-4309-b95d-6103408901ec" (UID: "5736742d-6a6c-4309-b95d-6103408901ec"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.285508 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovn-acl-logging/0.log" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.286113 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7b4bw_5736742d-6a6c-4309-b95d-6103408901ec/ovn-controller/0.log" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.286560 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" event={"ID":"5736742d-6a6c-4309-b95d-6103408901ec","Type":"ContainerDied","Data":"62df6ac04ca23854487e97a4cb353c94a24264ca3a1ab3705570624c599af8f6"} Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.286587 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7b4bw" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.286624 4954 scope.go:117] "RemoveContainer" containerID="03d8c619b3e5c8265fe9aee8335fefb2d829cc825e9b39932cf20879c1d29acc" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.287953 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" podUID="ab2760a1-3695-4556-bff0-d61d9da9d78b" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.288035 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" podUID="71312e1a-0fc4-417a-92e3-080c5651ff47" Dec 09 17:08:17 crc kubenswrapper[4954]: E1209 17:08:17.288275 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385\\\"\"" pod="openshift-operators/perses-operator-5446b9c989-2l8fr" podUID="2466d04c-561e-4c13-9939-9194d26ac50f" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.310469 4954 scope.go:117] "RemoveContainer" containerID="83cc6953e0379401e3f18419911b854006c6b0ef29c48c68913de23945108807" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.332301 4954 scope.go:117] "RemoveContainer" containerID="786c290f74c4753a5513172a7a5c61702fda44597b45a675d107b88fce0e7845" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.352029 4954 scope.go:117] "RemoveContainer" containerID="194b01a5ef3e0eedc093760d69ac51712d068f2a177b7ff585c79eb14749dbaf" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.364537 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-var-lib-openvswitch\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.364664 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65frk\" (UniqueName: \"kubernetes.io/projected/491c5049-7834-45f4-a9ac-8002941b4850-kube-api-access-65frk\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.364710 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/491c5049-7834-45f4-a9ac-8002941b4850-ovn-node-metrics-cert\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.364755 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-run-systemd\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.364890 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/491c5049-7834-45f4-a9ac-8002941b4850-env-overrides\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365034 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-slash\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365167 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/491c5049-7834-45f4-a9ac-8002941b4850-ovnkube-script-lib\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365216 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-run-ovn\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365238 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-run-ovn-kubernetes\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365294 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365327 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-run-netns\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365346 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-log-socket\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365364 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-cni-bin\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365386 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-run-openvswitch\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365482 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/491c5049-7834-45f4-a9ac-8002941b4850-ovnkube-config\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365500 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-systemd-units\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365541 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-kubelet\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365608 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-cni-netd\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365709 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-etc-openvswitch\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.365812 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-node-log\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.366017 4954 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5736742d-6a6c-4309-b95d-6103408901ec-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.366044 4954 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.366079 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msqm2\" (UniqueName: \"kubernetes.io/projected/5736742d-6a6c-4309-b95d-6103408901ec-kube-api-access-msqm2\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.366109 4954 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5736742d-6a6c-4309-b95d-6103408901ec-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.366124 4954 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5736742d-6a6c-4309-b95d-6103408901ec-host-slash\") on node \"crc\" DevicePath \"\"" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.374366 4954 scope.go:117] "RemoveContainer" containerID="2c2dcb2a286a0ee6e4cafac35abd757b814f6e678050c4a0f19b394333e20b0e" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.390564 4954 scope.go:117] "RemoveContainer" containerID="5dc35407ee180aeb01eabc7a10cf56e52d067687025bdb077fb6b7d11e2db8fb" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.396635 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7b4bw"] Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.412269 4954 scope.go:117] "RemoveContainer" containerID="933d452d327277f6d75b592ae77d9b5f58c66baa21be2d5eb6d48235ca96425e" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.412479 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7b4bw"] Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.424889 4954 scope.go:117] "RemoveContainer" containerID="503a479201e1f502015396b67d620ddc30efb9785ab0716d8216ebfe358d83e6" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.439919 4954 scope.go:117] "RemoveContainer" containerID="ec77437e8a005f33f917189a398fd7c01637c3afb9ec854646489738cc197981" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.467893 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-run-ovn\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.467954 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/491c5049-7834-45f4-a9ac-8002941b4850-ovnkube-script-lib\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.467993 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-run-ovn-kubernetes\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468040 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468075 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-run-netns\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468110 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-cni-bin\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468135 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-run-openvswitch\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468154 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-log-socket\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468188 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/491c5049-7834-45f4-a9ac-8002941b4850-ovnkube-config\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468212 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-systemd-units\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468235 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-kubelet\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468273 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-cni-netd\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468304 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-etc-openvswitch\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468335 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-node-log\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468375 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-var-lib-openvswitch\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468412 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65frk\" (UniqueName: \"kubernetes.io/projected/491c5049-7834-45f4-a9ac-8002941b4850-kube-api-access-65frk\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468453 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/491c5049-7834-45f4-a9ac-8002941b4850-ovn-node-metrics-cert\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468493 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-run-systemd\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468611 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/491c5049-7834-45f4-a9ac-8002941b4850-env-overrides\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468669 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-slash\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468786 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-slash\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.468850 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-run-ovn\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.469518 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-kubelet\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.469660 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-node-log\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.469603 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-cni-netd\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.469630 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-etc-openvswitch\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.469548 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-var-lib-openvswitch\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.469889 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/491c5049-7834-45f4-a9ac-8002941b4850-ovnkube-script-lib\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.469920 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-run-systemd\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.470270 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-run-ovn-kubernetes\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.470406 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/491c5049-7834-45f4-a9ac-8002941b4850-env-overrides\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.472377 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.479653 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/491c5049-7834-45f4-a9ac-8002941b4850-ovnkube-config\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.479730 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-run-openvswitch\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.479757 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-run-netns\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.479793 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-host-cni-bin\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.479826 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-log-socket\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.479871 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/491c5049-7834-45f4-a9ac-8002941b4850-systemd-units\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.510870 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/491c5049-7834-45f4-a9ac-8002941b4850-ovn-node-metrics-cert\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.529459 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65frk\" (UniqueName: \"kubernetes.io/projected/491c5049-7834-45f4-a9ac-8002941b4850-kube-api-access-65frk\") pod \"ovnkube-node-qjqcn\" (UID: \"491c5049-7834-45f4-a9ac-8002941b4850\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: I1209 17:08:17.556695 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:17 crc kubenswrapper[4954]: W1209 17:08:17.599436 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod491c5049_7834_45f4_a9ac_8002941b4850.slice/crio-31dba072b4b8edd0459e82cd5e60bd132a7dd8f93bbca0e961d8177072ca5732 WatchSource:0}: Error finding container 31dba072b4b8edd0459e82cd5e60bd132a7dd8f93bbca0e961d8177072ca5732: Status 404 returned error can't find the container with id 31dba072b4b8edd0459e82cd5e60bd132a7dd8f93bbca0e961d8177072ca5732 Dec 09 17:08:18 crc kubenswrapper[4954]: I1209 17:08:18.125889 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5736742d-6a6c-4309-b95d-6103408901ec" path="/var/lib/kubelet/pods/5736742d-6a6c-4309-b95d-6103408901ec/volumes" Dec 09 17:08:18 crc kubenswrapper[4954]: I1209 17:08:18.293860 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn" event={"ID":"0e70cc7a-1b64-47c6-b0f1-4da0e77c9870","Type":"ContainerStarted","Data":"168aa7b23bad87a0a376dc6b21af3263bdd5fb68433512254841b495bb595c80"} Dec 09 17:08:18 crc kubenswrapper[4954]: I1209 17:08:18.295951 4954 generic.go:334] "Generic (PLEG): container finished" podID="491c5049-7834-45f4-a9ac-8002941b4850" containerID="c8f7440f8b8ca7132cc63dcfb44fd9bbefc185d1df8f87eb41cc1119e0607e03" exitCode=0 Dec 09 17:08:18 crc kubenswrapper[4954]: I1209 17:08:18.295993 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" event={"ID":"491c5049-7834-45f4-a9ac-8002941b4850","Type":"ContainerDied","Data":"c8f7440f8b8ca7132cc63dcfb44fd9bbefc185d1df8f87eb41cc1119e0607e03"} Dec 09 17:08:18 crc kubenswrapper[4954]: I1209 17:08:18.296015 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" event={"ID":"491c5049-7834-45f4-a9ac-8002941b4850","Type":"ContainerStarted","Data":"31dba072b4b8edd0459e82cd5e60bd132a7dd8f93bbca0e961d8177072ca5732"} Dec 09 17:08:18 crc kubenswrapper[4954]: I1209 17:08:18.312907 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-gtwjn" podStartSLOduration=2.141898005 podStartE2EDuration="23.312890588s" podCreationTimestamp="2025-12-09 17:07:55 +0000 UTC" firstStartedPulling="2025-12-09 17:07:55.961124494 +0000 UTC m=+672.349298314" lastFinishedPulling="2025-12-09 17:08:17.132117077 +0000 UTC m=+693.520290897" observedRunningTime="2025-12-09 17:08:18.312007421 +0000 UTC m=+694.700181261" watchObservedRunningTime="2025-12-09 17:08:18.312890588 +0000 UTC m=+694.701064408" Dec 09 17:08:19 crc kubenswrapper[4954]: I1209 17:08:19.305259 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" event={"ID":"491c5049-7834-45f4-a9ac-8002941b4850","Type":"ContainerStarted","Data":"984d4ed931baa254301154531096e9a6994250cddc02cb10194a8d42cd340eb3"} Dec 09 17:08:19 crc kubenswrapper[4954]: I1209 17:08:19.305939 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" event={"ID":"491c5049-7834-45f4-a9ac-8002941b4850","Type":"ContainerStarted","Data":"a5e04bc5aab77db87f66a915196438b408705234ea70237027fe3a0b84a59eed"} Dec 09 17:08:19 crc kubenswrapper[4954]: I1209 17:08:19.305956 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" event={"ID":"491c5049-7834-45f4-a9ac-8002941b4850","Type":"ContainerStarted","Data":"3bf5eb4a8f0e3b32d72004ed663b8584698858800dd1520dff7835b58791f321"} Dec 09 17:08:19 crc kubenswrapper[4954]: I1209 17:08:19.305967 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" event={"ID":"491c5049-7834-45f4-a9ac-8002941b4850","Type":"ContainerStarted","Data":"a8449eddc928dfea790bb6d7a17d9751ebe5bcc5ea6a226c79b2c664d5c0b002"} Dec 09 17:08:19 crc kubenswrapper[4954]: I1209 17:08:19.305993 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" event={"ID":"491c5049-7834-45f4-a9ac-8002941b4850","Type":"ContainerStarted","Data":"1e56217de0bf25dc189a52fe8b3e6576cb348d0875abfe33def9e63a8f6befed"} Dec 09 17:08:20 crc kubenswrapper[4954]: I1209 17:08:20.315161 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" event={"ID":"491c5049-7834-45f4-a9ac-8002941b4850","Type":"ContainerStarted","Data":"ee6627fcf369c201e2437af3a8827c7eb9a56273afc9933eceb9bfc5febc7d04"} Dec 09 17:08:22 crc kubenswrapper[4954]: I1209 17:08:22.327048 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" event={"ID":"491c5049-7834-45f4-a9ac-8002941b4850","Type":"ContainerStarted","Data":"c00894e08383972029993ab2a9c44c776d182aed5ad7dec3a204701af011c397"} Dec 09 17:08:24 crc kubenswrapper[4954]: I1209 17:08:24.340810 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" event={"ID":"491c5049-7834-45f4-a9ac-8002941b4850","Type":"ContainerStarted","Data":"82c97da88d59f44d40a49a3317d8e9bd1bf1810b3da9c0a910f37ddeec29d350"} Dec 09 17:08:24 crc kubenswrapper[4954]: I1209 17:08:24.342439 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:24 crc kubenswrapper[4954]: I1209 17:08:24.342469 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:24 crc kubenswrapper[4954]: I1209 17:08:24.342512 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:24 crc kubenswrapper[4954]: I1209 17:08:24.372344 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:24 crc kubenswrapper[4954]: I1209 17:08:24.376571 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:24 crc kubenswrapper[4954]: I1209 17:08:24.382462 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" podStartSLOduration=7.3824519330000005 podStartE2EDuration="7.382451933s" podCreationTimestamp="2025-12-09 17:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:08:24.37660879 +0000 UTC m=+700.764782620" watchObservedRunningTime="2025-12-09 17:08:24.382451933 +0000 UTC m=+700.770625753" Dec 09 17:08:28 crc kubenswrapper[4954]: I1209 17:08:28.120029 4954 scope.go:117] "RemoveContainer" containerID="e0c232380ddfbb380b2f38c01d48754a185f0b2cee838436385d29e2945680e0" Dec 09 17:08:29 crc kubenswrapper[4954]: I1209 17:08:29.369570 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9n2h_1eccf2e0-30a3-4201-a23f-85a92a522d72/kube-multus/2.log" Dec 09 17:08:29 crc kubenswrapper[4954]: I1209 17:08:29.370215 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9n2h" event={"ID":"1eccf2e0-30a3-4201-a23f-85a92a522d72","Type":"ContainerStarted","Data":"17d3125e0bca4ef4df1b5a7038e015bb1d97dfa63756732fbd8a3c233ed3ec6a"} Dec 09 17:08:30 crc kubenswrapper[4954]: I1209 17:08:30.376659 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" event={"ID":"71312e1a-0fc4-417a-92e3-080c5651ff47","Type":"ContainerStarted","Data":"7041ce6112864440e3c7d4c5b5f62c67d4ffee44dffc6036a68b3234db315f6e"} Dec 09 17:08:30 crc kubenswrapper[4954]: I1209 17:08:30.378143 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" event={"ID":"ab2760a1-3695-4556-bff0-d61d9da9d78b","Type":"ContainerStarted","Data":"d529b134c9222a8941dc01f465152b51152586defbb48c71788e8aa975a5cac5"} Dec 09 17:08:30 crc kubenswrapper[4954]: I1209 17:08:30.414074 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn" podStartSLOduration=1.2552159 podStartE2EDuration="35.414054396s" podCreationTimestamp="2025-12-09 17:07:55 +0000 UTC" firstStartedPulling="2025-12-09 17:07:56.052229775 +0000 UTC m=+672.440403595" lastFinishedPulling="2025-12-09 17:08:30.211068271 +0000 UTC m=+706.599242091" observedRunningTime="2025-12-09 17:08:30.397216077 +0000 UTC m=+706.785389927" watchObservedRunningTime="2025-12-09 17:08:30.414054396 +0000 UTC m=+706.802228216" Dec 09 17:08:30 crc kubenswrapper[4954]: I1209 17:08:30.434925 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m" podStartSLOduration=1.845048491 podStartE2EDuration="35.43490317s" podCreationTimestamp="2025-12-09 17:07:55 +0000 UTC" firstStartedPulling="2025-12-09 17:07:56.099330584 +0000 UTC m=+672.487504404" lastFinishedPulling="2025-12-09 17:08:29.689185263 +0000 UTC m=+706.077359083" observedRunningTime="2025-12-09 17:08:30.428048156 +0000 UTC m=+706.816221976" watchObservedRunningTime="2025-12-09 17:08:30.43490317 +0000 UTC m=+706.823076990" Dec 09 17:08:31 crc kubenswrapper[4954]: I1209 17:08:31.384209 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" event={"ID":"cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6","Type":"ContainerStarted","Data":"504c60dcb6b6cbaee87f62959594c4f0909d084752f2f8391d60b18730be06aa"} Dec 09 17:08:31 crc kubenswrapper[4954]: I1209 17:08:31.384721 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" Dec 09 17:08:31 crc kubenswrapper[4954]: I1209 17:08:31.386742 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" Dec 09 17:08:31 crc kubenswrapper[4954]: I1209 17:08:31.406034 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-87ttl" podStartSLOduration=1.7535830780000001 podStartE2EDuration="36.406014206s" podCreationTimestamp="2025-12-09 17:07:55 +0000 UTC" firstStartedPulling="2025-12-09 17:07:56.273896025 +0000 UTC m=+672.662069845" lastFinishedPulling="2025-12-09 17:08:30.926327153 +0000 UTC m=+707.314500973" observedRunningTime="2025-12-09 17:08:31.402905179 +0000 UTC m=+707.791079009" watchObservedRunningTime="2025-12-09 17:08:31.406014206 +0000 UTC m=+707.794188026" Dec 09 17:08:34 crc kubenswrapper[4954]: I1209 17:08:34.401733 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-2l8fr" event={"ID":"2466d04c-561e-4c13-9939-9194d26ac50f","Type":"ContainerStarted","Data":"e293248341c100a6506e042296d0d2833ffa7971daf2a9b431fc9b1e5bcc8927"} Dec 09 17:08:34 crc kubenswrapper[4954]: I1209 17:08:34.402807 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-2l8fr" Dec 09 17:08:34 crc kubenswrapper[4954]: I1209 17:08:34.436575 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-2l8fr" podStartSLOduration=2.036832785 podStartE2EDuration="39.436548067s" podCreationTimestamp="2025-12-09 17:07:55 +0000 UTC" firstStartedPulling="2025-12-09 17:07:56.456951414 +0000 UTC m=+672.845125234" lastFinishedPulling="2025-12-09 17:08:33.856666696 +0000 UTC m=+710.244840516" observedRunningTime="2025-12-09 17:08:34.430096204 +0000 UTC m=+710.818270024" watchObservedRunningTime="2025-12-09 17:08:34.436548067 +0000 UTC m=+710.824721887" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.594745 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ltkfg"] Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.595822 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ltkfg" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.597840 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.605462 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pj9hg"] Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.606720 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-pj9hg" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.608973 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.609269 4954 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-qnftf" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.609725 4954 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-b6gq5" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.623712 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ltkfg"] Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.663216 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pj9hg"] Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.680305 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-lwjmk"] Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.681220 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-lwjmk" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.683822 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-lwjmk"] Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.685456 4954 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-nfkch" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.742215 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv5jf\" (UniqueName: \"kubernetes.io/projected/f09a1e60-44f9-4ce4-b136-ab1131433933-kube-api-access-nv5jf\") pod \"cert-manager-5b446d88c5-pj9hg\" (UID: \"f09a1e60-44f9-4ce4-b136-ab1131433933\") " pod="cert-manager/cert-manager-5b446d88c5-pj9hg" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.742282 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxwnm\" (UniqueName: \"kubernetes.io/projected/cec94679-b7b3-4210-9c39-a4f2c012266c-kube-api-access-dxwnm\") pod \"cert-manager-cainjector-7f985d654d-ltkfg\" (UID: \"cec94679-b7b3-4210-9c39-a4f2c012266c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ltkfg" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.843607 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv5jf\" (UniqueName: \"kubernetes.io/projected/f09a1e60-44f9-4ce4-b136-ab1131433933-kube-api-access-nv5jf\") pod \"cert-manager-5b446d88c5-pj9hg\" (UID: \"f09a1e60-44f9-4ce4-b136-ab1131433933\") " pod="cert-manager/cert-manager-5b446d88c5-pj9hg" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.843942 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k58v\" (UniqueName: \"kubernetes.io/projected/19d20d7b-da5f-446f-b2f8-cf6ab0864461-kube-api-access-6k58v\") pod \"cert-manager-webhook-5655c58dd6-lwjmk\" (UID: \"19d20d7b-da5f-446f-b2f8-cf6ab0864461\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-lwjmk" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.844042 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxwnm\" (UniqueName: \"kubernetes.io/projected/cec94679-b7b3-4210-9c39-a4f2c012266c-kube-api-access-dxwnm\") pod \"cert-manager-cainjector-7f985d654d-ltkfg\" (UID: \"cec94679-b7b3-4210-9c39-a4f2c012266c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ltkfg" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.863018 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxwnm\" (UniqueName: \"kubernetes.io/projected/cec94679-b7b3-4210-9c39-a4f2c012266c-kube-api-access-dxwnm\") pod \"cert-manager-cainjector-7f985d654d-ltkfg\" (UID: \"cec94679-b7b3-4210-9c39-a4f2c012266c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ltkfg" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.869703 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv5jf\" (UniqueName: \"kubernetes.io/projected/f09a1e60-44f9-4ce4-b136-ab1131433933-kube-api-access-nv5jf\") pod \"cert-manager-5b446d88c5-pj9hg\" (UID: \"f09a1e60-44f9-4ce4-b136-ab1131433933\") " pod="cert-manager/cert-manager-5b446d88c5-pj9hg" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.914395 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ltkfg" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.942516 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-pj9hg" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.944896 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k58v\" (UniqueName: \"kubernetes.io/projected/19d20d7b-da5f-446f-b2f8-cf6ab0864461-kube-api-access-6k58v\") pod \"cert-manager-webhook-5655c58dd6-lwjmk\" (UID: \"19d20d7b-da5f-446f-b2f8-cf6ab0864461\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-lwjmk" Dec 09 17:08:38 crc kubenswrapper[4954]: I1209 17:08:38.968459 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k58v\" (UniqueName: \"kubernetes.io/projected/19d20d7b-da5f-446f-b2f8-cf6ab0864461-kube-api-access-6k58v\") pod \"cert-manager-webhook-5655c58dd6-lwjmk\" (UID: \"19d20d7b-da5f-446f-b2f8-cf6ab0864461\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-lwjmk" Dec 09 17:08:39 crc kubenswrapper[4954]: I1209 17:08:39.003228 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-lwjmk" Dec 09 17:08:39 crc kubenswrapper[4954]: I1209 17:08:39.198520 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ltkfg"] Dec 09 17:08:39 crc kubenswrapper[4954]: W1209 17:08:39.214051 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcec94679_b7b3_4210_9c39_a4f2c012266c.slice/crio-2bcd50b27a805c057dec8f280fb07c2e988f90275e724a23499062d9dd48e384 WatchSource:0}: Error finding container 2bcd50b27a805c057dec8f280fb07c2e988f90275e724a23499062d9dd48e384: Status 404 returned error can't find the container with id 2bcd50b27a805c057dec8f280fb07c2e988f90275e724a23499062d9dd48e384 Dec 09 17:08:39 crc kubenswrapper[4954]: I1209 17:08:39.268746 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pj9hg"] Dec 09 17:08:39 crc kubenswrapper[4954]: W1209 17:08:39.285074 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf09a1e60_44f9_4ce4_b136_ab1131433933.slice/crio-c4e8eaf23cfd72ab1a3873ecad222f74c0a2e03460fbaee76fd6f4c2a079f957 WatchSource:0}: Error finding container c4e8eaf23cfd72ab1a3873ecad222f74c0a2e03460fbaee76fd6f4c2a079f957: Status 404 returned error can't find the container with id c4e8eaf23cfd72ab1a3873ecad222f74c0a2e03460fbaee76fd6f4c2a079f957 Dec 09 17:08:39 crc kubenswrapper[4954]: I1209 17:08:39.347213 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-lwjmk"] Dec 09 17:08:39 crc kubenswrapper[4954]: W1209 17:08:39.355513 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19d20d7b_da5f_446f_b2f8_cf6ab0864461.slice/crio-766329db780da1bed7584dc723a640f091027558064883f00b8c31b0c397dfb6 WatchSource:0}: Error finding container 766329db780da1bed7584dc723a640f091027558064883f00b8c31b0c397dfb6: Status 404 returned error can't find the container with id 766329db780da1bed7584dc723a640f091027558064883f00b8c31b0c397dfb6 Dec 09 17:08:39 crc kubenswrapper[4954]: I1209 17:08:39.428550 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ltkfg" event={"ID":"cec94679-b7b3-4210-9c39-a4f2c012266c","Type":"ContainerStarted","Data":"2bcd50b27a805c057dec8f280fb07c2e988f90275e724a23499062d9dd48e384"} Dec 09 17:08:39 crc kubenswrapper[4954]: I1209 17:08:39.429459 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-lwjmk" event={"ID":"19d20d7b-da5f-446f-b2f8-cf6ab0864461","Type":"ContainerStarted","Data":"766329db780da1bed7584dc723a640f091027558064883f00b8c31b0c397dfb6"} Dec 09 17:08:39 crc kubenswrapper[4954]: I1209 17:08:39.431058 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-pj9hg" event={"ID":"f09a1e60-44f9-4ce4-b136-ab1131433933","Type":"ContainerStarted","Data":"c4e8eaf23cfd72ab1a3873ecad222f74c0a2e03460fbaee76fd6f4c2a079f957"} Dec 09 17:08:46 crc kubenswrapper[4954]: I1209 17:08:46.072427 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-2l8fr" Dec 09 17:08:47 crc kubenswrapper[4954]: I1209 17:08:47.580312 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qjqcn" Dec 09 17:08:48 crc kubenswrapper[4954]: I1209 17:08:48.494791 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ltkfg" event={"ID":"cec94679-b7b3-4210-9c39-a4f2c012266c","Type":"ContainerStarted","Data":"e202f6c55fb58fe619abad9e4554d2cc69332c8b6e04e3f5761b3eea003da2f7"} Dec 09 17:08:48 crc kubenswrapper[4954]: I1209 17:08:48.497279 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-lwjmk" event={"ID":"19d20d7b-da5f-446f-b2f8-cf6ab0864461","Type":"ContainerStarted","Data":"7b4db2375e3efbb8a3149c3583549cc4860d55079687930be88e86867374e56b"} Dec 09 17:08:48 crc kubenswrapper[4954]: I1209 17:08:48.497425 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-lwjmk" Dec 09 17:08:48 crc kubenswrapper[4954]: I1209 17:08:48.498506 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-pj9hg" event={"ID":"f09a1e60-44f9-4ce4-b136-ab1131433933","Type":"ContainerStarted","Data":"e4be74d45a2b23a1d47bd68352a4131ae183cc337bb105b09392caf896dfaa9e"} Dec 09 17:08:48 crc kubenswrapper[4954]: I1209 17:08:48.545571 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-ltkfg" podStartSLOduration=2.069006965 podStartE2EDuration="10.545554349s" podCreationTimestamp="2025-12-09 17:08:38 +0000 UTC" firstStartedPulling="2025-12-09 17:08:39.223149882 +0000 UTC m=+715.611323702" lastFinishedPulling="2025-12-09 17:08:47.699697276 +0000 UTC m=+724.087871086" observedRunningTime="2025-12-09 17:08:48.544855187 +0000 UTC m=+724.933029017" watchObservedRunningTime="2025-12-09 17:08:48.545554349 +0000 UTC m=+724.933728169" Dec 09 17:08:48 crc kubenswrapper[4954]: I1209 17:08:48.565252 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-lwjmk" podStartSLOduration=2.311387811 podStartE2EDuration="10.565233495s" podCreationTimestamp="2025-12-09 17:08:38 +0000 UTC" firstStartedPulling="2025-12-09 17:08:39.357641962 +0000 UTC m=+715.745815782" lastFinishedPulling="2025-12-09 17:08:47.611487646 +0000 UTC m=+723.999661466" observedRunningTime="2025-12-09 17:08:48.561688564 +0000 UTC m=+724.949862384" watchObservedRunningTime="2025-12-09 17:08:48.565233495 +0000 UTC m=+724.953407305" Dec 09 17:08:48 crc kubenswrapper[4954]: I1209 17:08:48.607681 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-pj9hg" podStartSLOduration=2.290703164 podStartE2EDuration="10.607655743s" podCreationTimestamp="2025-12-09 17:08:38 +0000 UTC" firstStartedPulling="2025-12-09 17:08:39.294529337 +0000 UTC m=+715.682703157" lastFinishedPulling="2025-12-09 17:08:47.611481906 +0000 UTC m=+723.999655736" observedRunningTime="2025-12-09 17:08:48.602072928 +0000 UTC m=+724.990246758" watchObservedRunningTime="2025-12-09 17:08:48.607655743 +0000 UTC m=+724.995829563" Dec 09 17:08:54 crc kubenswrapper[4954]: I1209 17:08:54.006489 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-lwjmk" Dec 09 17:09:25 crc kubenswrapper[4954]: I1209 17:09:25.681569 4954 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.063930 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n"] Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.065631 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.068581 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.074197 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n"] Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.224210 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qcxh\" (UniqueName: \"kubernetes.io/projected/548caf7d-1620-4634-b87f-b7c6049386da-kube-api-access-5qcxh\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.224494 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.224798 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.250211 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz"] Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.252004 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.262668 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz"] Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.326616 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.326778 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qcxh\" (UniqueName: \"kubernetes.io/projected/548caf7d-1620-4634-b87f-b7c6049386da-kube-api-access-5qcxh\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.326824 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.327176 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.327350 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.354197 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qcxh\" (UniqueName: \"kubernetes.io/projected/548caf7d-1620-4634-b87f-b7c6049386da-kube-api-access-5qcxh\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.389359 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.429439 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.429582 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d582\" (UniqueName: \"kubernetes.io/projected/e59b591a-d66f-4c27-9151-fa15ff9f84cc-kube-api-access-6d582\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.429649 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.665794 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d582\" (UniqueName: \"kubernetes.io/projected/e59b591a-d66f-4c27-9151-fa15ff9f84cc-kube-api-access-6d582\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.665845 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.665895 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.666773 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.666845 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.700825 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d582\" (UniqueName: \"kubernetes.io/projected/e59b591a-d66f-4c27-9151-fa15ff9f84cc-kube-api-access-6d582\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.915019 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:26 crc kubenswrapper[4954]: I1209 17:09:26.991045 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n"] Dec 09 17:09:27 crc kubenswrapper[4954]: I1209 17:09:27.197664 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz"] Dec 09 17:09:27 crc kubenswrapper[4954]: W1209 17:09:27.206710 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode59b591a_d66f_4c27_9151_fa15ff9f84cc.slice/crio-3a86888c7e7d754c0f8d5f628bc97e4627372a84917ee10e9d356f501608c376 WatchSource:0}: Error finding container 3a86888c7e7d754c0f8d5f628bc97e4627372a84917ee10e9d356f501608c376: Status 404 returned error can't find the container with id 3a86888c7e7d754c0f8d5f628bc97e4627372a84917ee10e9d356f501608c376 Dec 09 17:09:27 crc kubenswrapper[4954]: I1209 17:09:27.760904 4954 generic.go:334] "Generic (PLEG): container finished" podID="548caf7d-1620-4634-b87f-b7c6049386da" containerID="a380e44d2794353a49b811e74d2a9168dc370fc2b3410331d23258074013c7cf" exitCode=0 Dec 09 17:09:27 crc kubenswrapper[4954]: I1209 17:09:27.761008 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" event={"ID":"548caf7d-1620-4634-b87f-b7c6049386da","Type":"ContainerDied","Data":"a380e44d2794353a49b811e74d2a9168dc370fc2b3410331d23258074013c7cf"} Dec 09 17:09:27 crc kubenswrapper[4954]: I1209 17:09:27.761034 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" event={"ID":"548caf7d-1620-4634-b87f-b7c6049386da","Type":"ContainerStarted","Data":"f533e5afa447c79fb14a188db5ea3a347fa7f31bf6602f9b291d4e18d28250a5"} Dec 09 17:09:27 crc kubenswrapper[4954]: I1209 17:09:27.762433 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" event={"ID":"e59b591a-d66f-4c27-9151-fa15ff9f84cc","Type":"ContainerStarted","Data":"57edee372489592207069d713a70d1b12993c4c614266b349c936e438ad99523"} Dec 09 17:09:27 crc kubenswrapper[4954]: I1209 17:09:27.762460 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" event={"ID":"e59b591a-d66f-4c27-9151-fa15ff9f84cc","Type":"ContainerStarted","Data":"3a86888c7e7d754c0f8d5f628bc97e4627372a84917ee10e9d356f501608c376"} Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.062042 4954 generic.go:334] "Generic (PLEG): container finished" podID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" containerID="57edee372489592207069d713a70d1b12993c4c614266b349c936e438ad99523" exitCode=0 Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.062523 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" event={"ID":"e59b591a-d66f-4c27-9151-fa15ff9f84cc","Type":"ContainerDied","Data":"57edee372489592207069d713a70d1b12993c4c614266b349c936e438ad99523"} Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.806428 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-djlr7"] Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.807864 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.821110 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djlr7"] Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.871916 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlh6d\" (UniqueName: \"kubernetes.io/projected/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-kube-api-access-tlh6d\") pod \"redhat-operators-djlr7\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.872011 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-utilities\") pod \"redhat-operators-djlr7\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.872048 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-catalog-content\") pod \"redhat-operators-djlr7\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.974100 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlh6d\" (UniqueName: \"kubernetes.io/projected/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-kube-api-access-tlh6d\") pod \"redhat-operators-djlr7\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.974681 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-utilities\") pod \"redhat-operators-djlr7\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.974715 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-catalog-content\") pod \"redhat-operators-djlr7\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.975263 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-catalog-content\") pod \"redhat-operators-djlr7\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.975208 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-utilities\") pod \"redhat-operators-djlr7\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:29 crc kubenswrapper[4954]: I1209 17:09:29.998424 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlh6d\" (UniqueName: \"kubernetes.io/projected/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-kube-api-access-tlh6d\") pod \"redhat-operators-djlr7\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:30 crc kubenswrapper[4954]: I1209 17:09:30.128435 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:30 crc kubenswrapper[4954]: I1209 17:09:30.996968 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djlr7"] Dec 09 17:09:31 crc kubenswrapper[4954]: I1209 17:09:31.077245 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djlr7" event={"ID":"7f1c5c55-d460-45e7-9307-2602d7bc6b8d","Type":"ContainerStarted","Data":"fe098672be8e438dd76aa747f93ccb541dc201293aa68df6922772833b55d401"} Dec 09 17:09:31 crc kubenswrapper[4954]: I1209 17:09:31.080229 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" event={"ID":"e59b591a-d66f-4c27-9151-fa15ff9f84cc","Type":"ContainerStarted","Data":"cfc1421f4b77cb3cb6ab534ac714be0d504d6433c614a0290944aafe2e3d75a4"} Dec 09 17:09:31 crc kubenswrapper[4954]: I1209 17:09:31.081742 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" event={"ID":"548caf7d-1620-4634-b87f-b7c6049386da","Type":"ContainerStarted","Data":"44b3a701ba31ae955f765ccd1957f59ceeb7e35d1f484b4529a86182192f8779"} Dec 09 17:09:32 crc kubenswrapper[4954]: I1209 17:09:32.091041 4954 generic.go:334] "Generic (PLEG): container finished" podID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" containerID="cfc1421f4b77cb3cb6ab534ac714be0d504d6433c614a0290944aafe2e3d75a4" exitCode=0 Dec 09 17:09:32 crc kubenswrapper[4954]: I1209 17:09:32.091128 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" event={"ID":"e59b591a-d66f-4c27-9151-fa15ff9f84cc","Type":"ContainerDied","Data":"cfc1421f4b77cb3cb6ab534ac714be0d504d6433c614a0290944aafe2e3d75a4"} Dec 09 17:09:32 crc kubenswrapper[4954]: I1209 17:09:32.093188 4954 generic.go:334] "Generic (PLEG): container finished" podID="548caf7d-1620-4634-b87f-b7c6049386da" containerID="44b3a701ba31ae955f765ccd1957f59ceeb7e35d1f484b4529a86182192f8779" exitCode=0 Dec 09 17:09:32 crc kubenswrapper[4954]: I1209 17:09:32.094274 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" event={"ID":"548caf7d-1620-4634-b87f-b7c6049386da","Type":"ContainerDied","Data":"44b3a701ba31ae955f765ccd1957f59ceeb7e35d1f484b4529a86182192f8779"} Dec 09 17:09:32 crc kubenswrapper[4954]: I1209 17:09:32.096537 4954 generic.go:334] "Generic (PLEG): container finished" podID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerID="6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b" exitCode=0 Dec 09 17:09:32 crc kubenswrapper[4954]: I1209 17:09:32.096578 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djlr7" event={"ID":"7f1c5c55-d460-45e7-9307-2602d7bc6b8d","Type":"ContainerDied","Data":"6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b"} Dec 09 17:09:33 crc kubenswrapper[4954]: I1209 17:09:33.108013 4954 generic.go:334] "Generic (PLEG): container finished" podID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" containerID="3404ff866c81781c3de706655cce7b77113eec6d7d280f18a431d3821b327c18" exitCode=0 Dec 09 17:09:33 crc kubenswrapper[4954]: I1209 17:09:33.108095 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" event={"ID":"e59b591a-d66f-4c27-9151-fa15ff9f84cc","Type":"ContainerDied","Data":"3404ff866c81781c3de706655cce7b77113eec6d7d280f18a431d3821b327c18"} Dec 09 17:09:33 crc kubenswrapper[4954]: I1209 17:09:33.111174 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" event={"ID":"548caf7d-1620-4634-b87f-b7c6049386da","Type":"ContainerStarted","Data":"d8767a284227046fd48b1de5c8cb57dfbb846c01b63269b6c76666feef584ec6"} Dec 09 17:09:33 crc kubenswrapper[4954]: I1209 17:09:33.148194 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" podStartSLOduration=4.299531343 podStartE2EDuration="7.148174904s" podCreationTimestamp="2025-12-09 17:09:26 +0000 UTC" firstStartedPulling="2025-12-09 17:09:27.763002699 +0000 UTC m=+764.151176519" lastFinishedPulling="2025-12-09 17:09:30.61164626 +0000 UTC m=+766.999820080" observedRunningTime="2025-12-09 17:09:33.146971967 +0000 UTC m=+769.535145797" watchObservedRunningTime="2025-12-09 17:09:33.148174904 +0000 UTC m=+769.536348724" Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.121895 4954 generic.go:334] "Generic (PLEG): container finished" podID="548caf7d-1620-4634-b87f-b7c6049386da" containerID="d8767a284227046fd48b1de5c8cb57dfbb846c01b63269b6c76666feef584ec6" exitCode=0 Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.142851 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djlr7" event={"ID":"7f1c5c55-d460-45e7-9307-2602d7bc6b8d","Type":"ContainerStarted","Data":"be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1"} Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.143267 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" event={"ID":"548caf7d-1620-4634-b87f-b7c6049386da","Type":"ContainerDied","Data":"d8767a284227046fd48b1de5c8cb57dfbb846c01b63269b6c76666feef584ec6"} Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.559468 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.645707 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d582\" (UniqueName: \"kubernetes.io/projected/e59b591a-d66f-4c27-9151-fa15ff9f84cc-kube-api-access-6d582\") pod \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.645782 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-bundle\") pod \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.645841 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-util\") pod \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\" (UID: \"e59b591a-d66f-4c27-9151-fa15ff9f84cc\") " Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.647153 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-bundle" (OuterVolumeSpecName: "bundle") pod "e59b591a-d66f-4c27-9151-fa15ff9f84cc" (UID: "e59b591a-d66f-4c27-9151-fa15ff9f84cc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.656385 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e59b591a-d66f-4c27-9151-fa15ff9f84cc-kube-api-access-6d582" (OuterVolumeSpecName: "kube-api-access-6d582") pod "e59b591a-d66f-4c27-9151-fa15ff9f84cc" (UID: "e59b591a-d66f-4c27-9151-fa15ff9f84cc"). InnerVolumeSpecName "kube-api-access-6d582". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.657157 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-util" (OuterVolumeSpecName: "util") pod "e59b591a-d66f-4c27-9151-fa15ff9f84cc" (UID: "e59b591a-d66f-4c27-9151-fa15ff9f84cc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.746456 4954 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-util\") on node \"crc\" DevicePath \"\"" Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.746483 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d582\" (UniqueName: \"kubernetes.io/projected/e59b591a-d66f-4c27-9151-fa15ff9f84cc-kube-api-access-6d582\") on node \"crc\" DevicePath \"\"" Dec 09 17:09:34 crc kubenswrapper[4954]: I1209 17:09:34.746495 4954 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e59b591a-d66f-4c27-9151-fa15ff9f84cc-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.130251 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.131290 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz" event={"ID":"e59b591a-d66f-4c27-9151-fa15ff9f84cc","Type":"ContainerDied","Data":"3a86888c7e7d754c0f8d5f628bc97e4627372a84917ee10e9d356f501608c376"} Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.131353 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a86888c7e7d754c0f8d5f628bc97e4627372a84917ee10e9d356f501608c376" Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.548104 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.558135 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qcxh\" (UniqueName: \"kubernetes.io/projected/548caf7d-1620-4634-b87f-b7c6049386da-kube-api-access-5qcxh\") pod \"548caf7d-1620-4634-b87f-b7c6049386da\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.558204 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-util\") pod \"548caf7d-1620-4634-b87f-b7c6049386da\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.558296 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-bundle\") pod \"548caf7d-1620-4634-b87f-b7c6049386da\" (UID: \"548caf7d-1620-4634-b87f-b7c6049386da\") " Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.560247 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-bundle" (OuterVolumeSpecName: "bundle") pod "548caf7d-1620-4634-b87f-b7c6049386da" (UID: "548caf7d-1620-4634-b87f-b7c6049386da"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.565070 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/548caf7d-1620-4634-b87f-b7c6049386da-kube-api-access-5qcxh" (OuterVolumeSpecName: "kube-api-access-5qcxh") pod "548caf7d-1620-4634-b87f-b7c6049386da" (UID: "548caf7d-1620-4634-b87f-b7c6049386da"). InnerVolumeSpecName "kube-api-access-5qcxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.574934 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-util" (OuterVolumeSpecName: "util") pod "548caf7d-1620-4634-b87f-b7c6049386da" (UID: "548caf7d-1620-4634-b87f-b7c6049386da"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.660518 4954 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-util\") on node \"crc\" DevicePath \"\"" Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.660653 4954 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/548caf7d-1620-4634-b87f-b7c6049386da-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:09:35 crc kubenswrapper[4954]: I1209 17:09:35.660665 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qcxh\" (UniqueName: \"kubernetes.io/projected/548caf7d-1620-4634-b87f-b7c6049386da-kube-api-access-5qcxh\") on node \"crc\" DevicePath \"\"" Dec 09 17:09:36 crc kubenswrapper[4954]: I1209 17:09:36.141602 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" event={"ID":"548caf7d-1620-4634-b87f-b7c6049386da","Type":"ContainerDied","Data":"f533e5afa447c79fb14a188db5ea3a347fa7f31bf6602f9b291d4e18d28250a5"} Dec 09 17:09:36 crc kubenswrapper[4954]: I1209 17:09:36.141972 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f533e5afa447c79fb14a188db5ea3a347fa7f31bf6602f9b291d4e18d28250a5" Dec 09 17:09:36 crc kubenswrapper[4954]: I1209 17:09:36.141867 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n" Dec 09 17:09:36 crc kubenswrapper[4954]: I1209 17:09:36.143133 4954 generic.go:334] "Generic (PLEG): container finished" podID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerID="be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1" exitCode=0 Dec 09 17:09:36 crc kubenswrapper[4954]: I1209 17:09:36.143165 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djlr7" event={"ID":"7f1c5c55-d460-45e7-9307-2602d7bc6b8d","Type":"ContainerDied","Data":"be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1"} Dec 09 17:09:37 crc kubenswrapper[4954]: I1209 17:09:37.151096 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djlr7" event={"ID":"7f1c5c55-d460-45e7-9307-2602d7bc6b8d","Type":"ContainerStarted","Data":"357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5"} Dec 09 17:09:37 crc kubenswrapper[4954]: I1209 17:09:37.168740 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-djlr7" podStartSLOduration=3.616277748 podStartE2EDuration="8.168724663s" podCreationTimestamp="2025-12-09 17:09:29 +0000 UTC" firstStartedPulling="2025-12-09 17:09:32.105224164 +0000 UTC m=+768.493397984" lastFinishedPulling="2025-12-09 17:09:36.657671079 +0000 UTC m=+773.045844899" observedRunningTime="2025-12-09 17:09:37.167960688 +0000 UTC m=+773.556134508" watchObservedRunningTime="2025-12-09 17:09:37.168724663 +0000 UTC m=+773.556898483" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.599294 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-c7nbc"] Dec 09 17:09:39 crc kubenswrapper[4954]: E1209 17:09:39.599973 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548caf7d-1620-4634-b87f-b7c6049386da" containerName="pull" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.599998 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="548caf7d-1620-4634-b87f-b7c6049386da" containerName="pull" Dec 09 17:09:39 crc kubenswrapper[4954]: E1209 17:09:39.600019 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" containerName="util" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.600030 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" containerName="util" Dec 09 17:09:39 crc kubenswrapper[4954]: E1209 17:09:39.600049 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" containerName="pull" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.600606 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" containerName="pull" Dec 09 17:09:39 crc kubenswrapper[4954]: E1209 17:09:39.600628 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" containerName="extract" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.600638 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" containerName="extract" Dec 09 17:09:39 crc kubenswrapper[4954]: E1209 17:09:39.600667 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548caf7d-1620-4634-b87f-b7c6049386da" containerName="extract" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.600675 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="548caf7d-1620-4634-b87f-b7c6049386da" containerName="extract" Dec 09 17:09:39 crc kubenswrapper[4954]: E1209 17:09:39.600703 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548caf7d-1620-4634-b87f-b7c6049386da" containerName="util" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.600710 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="548caf7d-1620-4634-b87f-b7c6049386da" containerName="util" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.611873 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" containerName="extract" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.611953 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="548caf7d-1620-4634-b87f-b7c6049386da" containerName="extract" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.612948 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-c7nbc" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.616293 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.616773 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-vzvhf" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.617399 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.644616 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-c7nbc"] Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.713519 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szkcf\" (UniqueName: \"kubernetes.io/projected/0f07fc00-40fe-4979-9225-d66d3c359c2c-kube-api-access-szkcf\") pod \"cluster-logging-operator-ff9846bd-c7nbc\" (UID: \"0f07fc00-40fe-4979-9225-d66d3c359c2c\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-c7nbc" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.815651 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szkcf\" (UniqueName: \"kubernetes.io/projected/0f07fc00-40fe-4979-9225-d66d3c359c2c-kube-api-access-szkcf\") pod \"cluster-logging-operator-ff9846bd-c7nbc\" (UID: \"0f07fc00-40fe-4979-9225-d66d3c359c2c\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-c7nbc" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.837954 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szkcf\" (UniqueName: \"kubernetes.io/projected/0f07fc00-40fe-4979-9225-d66d3c359c2c-kube-api-access-szkcf\") pod \"cluster-logging-operator-ff9846bd-c7nbc\" (UID: \"0f07fc00-40fe-4979-9225-d66d3c359c2c\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-c7nbc" Dec 09 17:09:39 crc kubenswrapper[4954]: I1209 17:09:39.963509 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-c7nbc" Dec 09 17:09:40 crc kubenswrapper[4954]: I1209 17:09:40.152885 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:40 crc kubenswrapper[4954]: I1209 17:09:40.153414 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:40 crc kubenswrapper[4954]: I1209 17:09:40.237025 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-c7nbc"] Dec 09 17:09:40 crc kubenswrapper[4954]: W1209 17:09:40.253761 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f07fc00_40fe_4979_9225_d66d3c359c2c.slice/crio-e44d5c0b1a3fa7eebc3ceb3f788d674e076444dd5178862c19a4a6b7eed58d9a WatchSource:0}: Error finding container e44d5c0b1a3fa7eebc3ceb3f788d674e076444dd5178862c19a4a6b7eed58d9a: Status 404 returned error can't find the container with id e44d5c0b1a3fa7eebc3ceb3f788d674e076444dd5178862c19a4a6b7eed58d9a Dec 09 17:09:41 crc kubenswrapper[4954]: I1209 17:09:41.189138 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-c7nbc" event={"ID":"0f07fc00-40fe-4979-9225-d66d3c359c2c","Type":"ContainerStarted","Data":"e44d5c0b1a3fa7eebc3ceb3f788d674e076444dd5178862c19a4a6b7eed58d9a"} Dec 09 17:09:41 crc kubenswrapper[4954]: I1209 17:09:41.214210 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-djlr7" podUID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerName="registry-server" probeResult="failure" output=< Dec 09 17:09:41 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 17:09:41 crc kubenswrapper[4954]: > Dec 09 17:09:43 crc kubenswrapper[4954]: I1209 17:09:43.755838 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:09:43 crc kubenswrapper[4954]: I1209 17:09:43.756225 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.533243 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95"] Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.534946 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: W1209 17:09:49.537643 4954 reflector.go:561] object-"openshift-operators-redhat"/"loki-operator-manager-config": failed to list *v1.ConfigMap: configmaps "loki-operator-manager-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-operators-redhat": no relationship found between node 'crc' and this object Dec 09 17:09:49 crc kubenswrapper[4954]: E1209 17:09:49.537723 4954 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators-redhat\"/\"loki-operator-manager-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"loki-operator-manager-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-operators-redhat\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.539319 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.541281 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-mt6td" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.541355 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.541511 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.546702 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.576483 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95"] Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.684860 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/73f81f6d-15c4-4cc8-95be-59c4db2139ea-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.684942 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73f81f6d-15c4-4cc8-95be-59c4db2139ea-webhook-cert\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.684985 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/73f81f6d-15c4-4cc8-95be-59c4db2139ea-manager-config\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.685024 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73f81f6d-15c4-4cc8-95be-59c4db2139ea-apiservice-cert\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.685057 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwtww\" (UniqueName: \"kubernetes.io/projected/73f81f6d-15c4-4cc8-95be-59c4db2139ea-kube-api-access-mwtww\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.786259 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/73f81f6d-15c4-4cc8-95be-59c4db2139ea-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.786375 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73f81f6d-15c4-4cc8-95be-59c4db2139ea-webhook-cert\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.786423 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/73f81f6d-15c4-4cc8-95be-59c4db2139ea-manager-config\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.786459 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73f81f6d-15c4-4cc8-95be-59c4db2139ea-apiservice-cert\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.786491 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwtww\" (UniqueName: \"kubernetes.io/projected/73f81f6d-15c4-4cc8-95be-59c4db2139ea-kube-api-access-mwtww\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.793980 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/73f81f6d-15c4-4cc8-95be-59c4db2139ea-webhook-cert\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.794250 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/73f81f6d-15c4-4cc8-95be-59c4db2139ea-apiservice-cert\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.807662 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/73f81f6d-15c4-4cc8-95be-59c4db2139ea-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:49 crc kubenswrapper[4954]: I1209 17:09:49.853527 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwtww\" (UniqueName: \"kubernetes.io/projected/73f81f6d-15c4-4cc8-95be-59c4db2139ea-kube-api-access-mwtww\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:50 crc kubenswrapper[4954]: I1209 17:09:50.353021 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:50 crc kubenswrapper[4954]: I1209 17:09:50.460004 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 09 17:09:50 crc kubenswrapper[4954]: I1209 17:09:50.467751 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/73f81f6d-15c4-4cc8-95be-59c4db2139ea-manager-config\") pod \"loki-operator-controller-manager-5f76df48c6-fvg95\" (UID: \"73f81f6d-15c4-4cc8-95be-59c4db2139ea\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:50 crc kubenswrapper[4954]: I1209 17:09:50.643304 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:50 crc kubenswrapper[4954]: I1209 17:09:50.755213 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:09:52 crc kubenswrapper[4954]: I1209 17:09:52.995898 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-djlr7"] Dec 09 17:09:52 crc kubenswrapper[4954]: I1209 17:09:52.996440 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-djlr7" podUID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerName="registry-server" containerID="cri-o://357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5" gracePeriod=2 Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.416897 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.447429 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95"] Dec 09 17:09:54 crc kubenswrapper[4954]: W1209 17:09:54.456477 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73f81f6d_15c4_4cc8_95be_59c4db2139ea.slice/crio-439d3354029ddd0a612c67c6c73b7fa74dc9ad4ed06e79a2133caaebe37abc98 WatchSource:0}: Error finding container 439d3354029ddd0a612c67c6c73b7fa74dc9ad4ed06e79a2133caaebe37abc98: Status 404 returned error can't find the container with id 439d3354029ddd0a612c67c6c73b7fa74dc9ad4ed06e79a2133caaebe37abc98 Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.457489 4954 generic.go:334] "Generic (PLEG): container finished" podID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerID="357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5" exitCode=0 Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.457612 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djlr7" event={"ID":"7f1c5c55-d460-45e7-9307-2602d7bc6b8d","Type":"ContainerDied","Data":"357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5"} Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.457683 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djlr7" event={"ID":"7f1c5c55-d460-45e7-9307-2602d7bc6b8d","Type":"ContainerDied","Data":"fe098672be8e438dd76aa747f93ccb541dc201293aa68df6922772833b55d401"} Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.457636 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djlr7" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.457705 4954 scope.go:117] "RemoveContainer" containerID="357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.465389 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-c7nbc" event={"ID":"0f07fc00-40fe-4979-9225-d66d3c359c2c","Type":"ContainerStarted","Data":"441a9d6b3212f1331df84b07f8af4448781892fa1814621747c48f431da79552"} Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.483062 4954 scope.go:117] "RemoveContainer" containerID="be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.492760 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-c7nbc" podStartSLOduration=1.642078575 podStartE2EDuration="15.49272621s" podCreationTimestamp="2025-12-09 17:09:39 +0000 UTC" firstStartedPulling="2025-12-09 17:09:40.257769988 +0000 UTC m=+776.645943808" lastFinishedPulling="2025-12-09 17:09:54.108417623 +0000 UTC m=+790.496591443" observedRunningTime="2025-12-09 17:09:54.491067519 +0000 UTC m=+790.879241349" watchObservedRunningTime="2025-12-09 17:09:54.49272621 +0000 UTC m=+790.880900030" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.522430 4954 scope.go:117] "RemoveContainer" containerID="6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.537561 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlh6d\" (UniqueName: \"kubernetes.io/projected/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-kube-api-access-tlh6d\") pod \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.537773 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-catalog-content\") pod \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.537823 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-utilities\") pod \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\" (UID: \"7f1c5c55-d460-45e7-9307-2602d7bc6b8d\") " Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.540489 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-utilities" (OuterVolumeSpecName: "utilities") pod "7f1c5c55-d460-45e7-9307-2602d7bc6b8d" (UID: "7f1c5c55-d460-45e7-9307-2602d7bc6b8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.552422 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-kube-api-access-tlh6d" (OuterVolumeSpecName: "kube-api-access-tlh6d") pod "7f1c5c55-d460-45e7-9307-2602d7bc6b8d" (UID: "7f1c5c55-d460-45e7-9307-2602d7bc6b8d"). InnerVolumeSpecName "kube-api-access-tlh6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.579770 4954 scope.go:117] "RemoveContainer" containerID="357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5" Dec 09 17:09:54 crc kubenswrapper[4954]: E1209 17:09:54.581447 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5\": container with ID starting with 357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5 not found: ID does not exist" containerID="357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.581501 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5"} err="failed to get container status \"357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5\": rpc error: code = NotFound desc = could not find container \"357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5\": container with ID starting with 357e3604a6dfe7a74ec78bddb6e7388631a1b5243d1451d9ec50df28459d8fe5 not found: ID does not exist" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.581530 4954 scope.go:117] "RemoveContainer" containerID="be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1" Dec 09 17:09:54 crc kubenswrapper[4954]: E1209 17:09:54.584178 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1\": container with ID starting with be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1 not found: ID does not exist" containerID="be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.584219 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1"} err="failed to get container status \"be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1\": rpc error: code = NotFound desc = could not find container \"be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1\": container with ID starting with be04b809d43c79b894889328e21ab383dcf0219d56e54caf84cb3a68cdd0f5d1 not found: ID does not exist" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.584242 4954 scope.go:117] "RemoveContainer" containerID="6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b" Dec 09 17:09:54 crc kubenswrapper[4954]: E1209 17:09:54.584520 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b\": container with ID starting with 6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b not found: ID does not exist" containerID="6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.584544 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b"} err="failed to get container status \"6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b\": rpc error: code = NotFound desc = could not find container \"6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b\": container with ID starting with 6e35c0baf174ec11d6084684cba6346e2e007932e045647d2144ab6724e1c21b not found: ID does not exist" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.639179 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.639210 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlh6d\" (UniqueName: \"kubernetes.io/projected/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-kube-api-access-tlh6d\") on node \"crc\" DevicePath \"\"" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.698305 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f1c5c55-d460-45e7-9307-2602d7bc6b8d" (UID: "7f1c5c55-d460-45e7-9307-2602d7bc6b8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.740275 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1c5c55-d460-45e7-9307-2602d7bc6b8d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.791500 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-djlr7"] Dec 09 17:09:54 crc kubenswrapper[4954]: I1209 17:09:54.796744 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-djlr7"] Dec 09 17:09:55 crc kubenswrapper[4954]: I1209 17:09:55.481425 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" event={"ID":"73f81f6d-15c4-4cc8-95be-59c4db2139ea","Type":"ContainerStarted","Data":"439d3354029ddd0a612c67c6c73b7fa74dc9ad4ed06e79a2133caaebe37abc98"} Dec 09 17:09:56 crc kubenswrapper[4954]: I1209 17:09:56.130743 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" path="/var/lib/kubelet/pods/7f1c5c55-d460-45e7-9307-2602d7bc6b8d/volumes" Dec 09 17:09:58 crc kubenswrapper[4954]: I1209 17:09:58.503705 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" event={"ID":"73f81f6d-15c4-4cc8-95be-59c4db2139ea","Type":"ContainerStarted","Data":"447aeb4cef1d1befc619f22e496d1b73fbc067e003a21fdf68fe4ab938e0f738"} Dec 09 17:10:05 crc kubenswrapper[4954]: I1209 17:10:05.974758 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" event={"ID":"73f81f6d-15c4-4cc8-95be-59c4db2139ea","Type":"ContainerStarted","Data":"bfe1868b2815c39267f932d0b86fa8bed9c0b2390988979106f4dd68c365844a"} Dec 09 17:10:05 crc kubenswrapper[4954]: I1209 17:10:05.975566 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:10:05 crc kubenswrapper[4954]: I1209 17:10:05.978628 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" Dec 09 17:10:06 crc kubenswrapper[4954]: I1209 17:10:06.008083 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-5f76df48c6-fvg95" podStartSLOduration=6.082099835 podStartE2EDuration="17.008064118s" podCreationTimestamp="2025-12-09 17:09:49 +0000 UTC" firstStartedPulling="2025-12-09 17:09:54.461356548 +0000 UTC m=+790.849530368" lastFinishedPulling="2025-12-09 17:10:05.387320821 +0000 UTC m=+801.775494651" observedRunningTime="2025-12-09 17:10:06.006445987 +0000 UTC m=+802.394619817" watchObservedRunningTime="2025-12-09 17:10:06.008064118 +0000 UTC m=+802.396237938" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.760323 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 09 17:10:09 crc kubenswrapper[4954]: E1209 17:10:09.761196 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerName="extract-content" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.761213 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerName="extract-content" Dec 09 17:10:09 crc kubenswrapper[4954]: E1209 17:10:09.761228 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerName="extract-utilities" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.761236 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerName="extract-utilities" Dec 09 17:10:09 crc kubenswrapper[4954]: E1209 17:10:09.761245 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerName="registry-server" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.761256 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerName="registry-server" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.761391 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f1c5c55-d460-45e7-9307-2602d7bc6b8d" containerName="registry-server" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.761854 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.765378 4954 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-n6l8x" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.766612 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.766638 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.786032 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.934754 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-77c04501-3229-4f86-ab45-8bfd6e1ad86d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77c04501-3229-4f86-ab45-8bfd6e1ad86d\") pod \"minio\" (UID: \"014b33f0-9e86-4533-b1f7-0d3293423a5e\") " pod="minio-dev/minio" Dec 09 17:10:09 crc kubenswrapper[4954]: I1209 17:10:09.934815 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxzfh\" (UniqueName: \"kubernetes.io/projected/014b33f0-9e86-4533-b1f7-0d3293423a5e-kube-api-access-vxzfh\") pod \"minio\" (UID: \"014b33f0-9e86-4533-b1f7-0d3293423a5e\") " pod="minio-dev/minio" Dec 09 17:10:10 crc kubenswrapper[4954]: I1209 17:10:10.036090 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-77c04501-3229-4f86-ab45-8bfd6e1ad86d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77c04501-3229-4f86-ab45-8bfd6e1ad86d\") pod \"minio\" (UID: \"014b33f0-9e86-4533-b1f7-0d3293423a5e\") " pod="minio-dev/minio" Dec 09 17:10:10 crc kubenswrapper[4954]: I1209 17:10:10.036164 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxzfh\" (UniqueName: \"kubernetes.io/projected/014b33f0-9e86-4533-b1f7-0d3293423a5e-kube-api-access-vxzfh\") pod \"minio\" (UID: \"014b33f0-9e86-4533-b1f7-0d3293423a5e\") " pod="minio-dev/minio" Dec 09 17:10:10 crc kubenswrapper[4954]: I1209 17:10:10.040092 4954 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 17:10:10 crc kubenswrapper[4954]: I1209 17:10:10.040165 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-77c04501-3229-4f86-ab45-8bfd6e1ad86d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77c04501-3229-4f86-ab45-8bfd6e1ad86d\") pod \"minio\" (UID: \"014b33f0-9e86-4533-b1f7-0d3293423a5e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fa6bdfd24ca2ea2ca3c12f2585387ba69ca681b72362999b98b3a5686ebb2791/globalmount\"" pod="minio-dev/minio" Dec 09 17:10:10 crc kubenswrapper[4954]: I1209 17:10:10.063489 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxzfh\" (UniqueName: \"kubernetes.io/projected/014b33f0-9e86-4533-b1f7-0d3293423a5e-kube-api-access-vxzfh\") pod \"minio\" (UID: \"014b33f0-9e86-4533-b1f7-0d3293423a5e\") " pod="minio-dev/minio" Dec 09 17:10:10 crc kubenswrapper[4954]: I1209 17:10:10.085889 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-77c04501-3229-4f86-ab45-8bfd6e1ad86d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77c04501-3229-4f86-ab45-8bfd6e1ad86d\") pod \"minio\" (UID: \"014b33f0-9e86-4533-b1f7-0d3293423a5e\") " pod="minio-dev/minio" Dec 09 17:10:10 crc kubenswrapper[4954]: I1209 17:10:10.381584 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 09 17:10:10 crc kubenswrapper[4954]: I1209 17:10:10.664491 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 09 17:10:11 crc kubenswrapper[4954]: I1209 17:10:11.004433 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"014b33f0-9e86-4533-b1f7-0d3293423a5e","Type":"ContainerStarted","Data":"e8ce75309766c737a136384bd4e098b15a3b6413734ebb8e0941f2cffbe4e17e"} Dec 09 17:10:13 crc kubenswrapper[4954]: I1209 17:10:13.755085 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:10:13 crc kubenswrapper[4954]: I1209 17:10:13.755407 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:10:16 crc kubenswrapper[4954]: I1209 17:10:16.040456 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"014b33f0-9e86-4533-b1f7-0d3293423a5e","Type":"ContainerStarted","Data":"0d75a01433b7a4dd365b57962887fe11916a2df9a562cf42074afe433fed9325"} Dec 09 17:10:16 crc kubenswrapper[4954]: I1209 17:10:16.057151 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.356048889 podStartE2EDuration="9.057127116s" podCreationTimestamp="2025-12-09 17:10:07 +0000 UTC" firstStartedPulling="2025-12-09 17:10:10.680071444 +0000 UTC m=+807.068245264" lastFinishedPulling="2025-12-09 17:10:15.381149671 +0000 UTC m=+811.769323491" observedRunningTime="2025-12-09 17:10:16.057077265 +0000 UTC m=+812.445251085" watchObservedRunningTime="2025-12-09 17:10:16.057127116 +0000 UTC m=+812.445300936" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.344372 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr"] Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.347127 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.357318 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.357566 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.357738 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.357921 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-kjh2r" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.358077 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.371823 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr"] Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.490702 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.490860 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.490969 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.491127 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-config\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.491226 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdt7z\" (UniqueName: \"kubernetes.io/projected/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-kube-api-access-jdt7z\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.518094 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-stgnv"] Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.518825 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.527177 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.527297 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.528510 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.531953 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-stgnv"] Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.592959 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.593004 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.593040 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.593062 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.593097 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.593115 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7dd2098-b496-49e7-9cdb-f87431dc135f-config\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.593135 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-config\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.593169 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdt7z\" (UniqueName: \"kubernetes.io/projected/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-kube-api-access-jdt7z\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.593191 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zksq\" (UniqueName: \"kubernetes.io/projected/f7dd2098-b496-49e7-9cdb-f87431dc135f-kube-api-access-9zksq\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.593207 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.593231 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.594994 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.595371 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-config\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.605800 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.606377 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.614776 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd"] Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.615974 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.620157 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.620401 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.637138 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdt7z\" (UniqueName: \"kubernetes.io/projected/8ca5d9e5-51f2-4a7a-86e4-a862190a0828-kube-api-access-jdt7z\") pod \"logging-loki-distributor-76cc67bf56-hgwvr\" (UID: \"8ca5d9e5-51f2-4a7a-86e4-a862190a0828\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.642702 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd"] Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.682205 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695041 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695091 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695124 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw6ct\" (UniqueName: \"kubernetes.io/projected/73361aee-2741-4328-b840-bcc3619f8cc0-kube-api-access-kw6ct\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695165 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695187 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7dd2098-b496-49e7-9cdb-f87431dc135f-config\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695205 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73361aee-2741-4328-b840-bcc3619f8cc0-config\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695220 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/73361aee-2741-4328-b840-bcc3619f8cc0-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695244 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73361aee-2741-4328-b840-bcc3619f8cc0-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695264 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/73361aee-2741-4328-b840-bcc3619f8cc0-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695292 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zksq\" (UniqueName: \"kubernetes.io/projected/f7dd2098-b496-49e7-9cdb-f87431dc135f-kube-api-access-9zksq\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.695310 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.696212 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.700195 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.701396 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7dd2098-b496-49e7-9cdb-f87431dc135f-config\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.702051 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.702140 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/f7dd2098-b496-49e7-9cdb-f87431dc135f-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.730456 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zksq\" (UniqueName: \"kubernetes.io/projected/f7dd2098-b496-49e7-9cdb-f87431dc135f-kube-api-access-9zksq\") pod \"logging-loki-querier-5895d59bb8-stgnv\" (UID: \"f7dd2098-b496-49e7-9cdb-f87431dc135f\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.752318 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-676969bccf-mxr7n"] Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.756067 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.760947 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.761292 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.761474 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.761685 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.761880 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.790937 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-676969bccf-hqhwj"] Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.792251 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.794163 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-676969bccf-mxr7n"] Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.795296 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-sg5mr" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.796056 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/73361aee-2741-4328-b840-bcc3619f8cc0-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.796185 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw6ct\" (UniqueName: \"kubernetes.io/projected/73361aee-2741-4328-b840-bcc3619f8cc0-kube-api-access-kw6ct\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.796225 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73361aee-2741-4328-b840-bcc3619f8cc0-config\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.796248 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/73361aee-2741-4328-b840-bcc3619f8cc0-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.796279 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73361aee-2741-4328-b840-bcc3619f8cc0-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.797254 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/73361aee-2741-4328-b840-bcc3619f8cc0-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.797757 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73361aee-2741-4328-b840-bcc3619f8cc0-config\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.810894 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/73361aee-2741-4328-b840-bcc3619f8cc0-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.814446 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/73361aee-2741-4328-b840-bcc3619f8cc0-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.833604 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-676969bccf-hqhwj"] Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.842180 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:20 crc kubenswrapper[4954]: I1209 17:10:20.845384 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw6ct\" (UniqueName: \"kubernetes.io/projected/73361aee-2741-4328-b840-bcc3619f8cc0-kube-api-access-kw6ct\") pod \"logging-loki-query-frontend-84558f7c9f-gzndd\" (UID: \"73361aee-2741-4328-b840-bcc3619f8cc0\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.003975 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005014 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5qdf\" (UniqueName: \"kubernetes.io/projected/49d8d406-58db-4226-806b-850e6e12b04f-kube-api-access-l5qdf\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005088 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-lokistack-gateway\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005120 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005153 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/81042cee-7dc6-4fc9-b8cf-70a4eff73510-tls-secret\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005174 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-logging-loki-ca-bundle\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005195 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhttz\" (UniqueName: \"kubernetes.io/projected/81042cee-7dc6-4fc9-b8cf-70a4eff73510-kube-api-access-fhttz\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005215 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/81042cee-7dc6-4fc9-b8cf-70a4eff73510-tenants\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005235 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-rbac\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005254 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-logging-loki-ca-bundle\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005282 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-lokistack-gateway\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005307 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/81042cee-7dc6-4fc9-b8cf-70a4eff73510-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005453 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-rbac\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005492 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/49d8d406-58db-4226-806b-850e6e12b04f-tenants\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005524 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/49d8d406-58db-4226-806b-850e6e12b04f-tls-secret\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005551 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/49d8d406-58db-4226-806b-850e6e12b04f-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.005580 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.106902 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-lokistack-gateway\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.107515 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/81042cee-7dc6-4fc9-b8cf-70a4eff73510-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.107583 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-rbac\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.107647 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/49d8d406-58db-4226-806b-850e6e12b04f-tenants\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.107686 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/49d8d406-58db-4226-806b-850e6e12b04f-tls-secret\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.107714 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/49d8d406-58db-4226-806b-850e6e12b04f-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.107751 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.107833 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5qdf\" (UniqueName: \"kubernetes.io/projected/49d8d406-58db-4226-806b-850e6e12b04f-kube-api-access-l5qdf\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.107887 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-lokistack-gateway\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.107933 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.108006 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/81042cee-7dc6-4fc9-b8cf-70a4eff73510-tls-secret\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.108050 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-logging-loki-ca-bundle\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.108090 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhttz\" (UniqueName: \"kubernetes.io/projected/81042cee-7dc6-4fc9-b8cf-70a4eff73510-kube-api-access-fhttz\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.108132 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/81042cee-7dc6-4fc9-b8cf-70a4eff73510-tenants\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.108184 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-rbac\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.108232 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-logging-loki-ca-bundle\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.109755 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-logging-loki-ca-bundle\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.112162 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-lokistack-gateway\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.112792 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-lokistack-gateway\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.115006 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.120263 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/81042cee-7dc6-4fc9-b8cf-70a4eff73510-tls-secret\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.121366 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-logging-loki-ca-bundle\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.123870 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/49d8d406-58db-4226-806b-850e6e12b04f-rbac\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.133136 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.134331 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/81042cee-7dc6-4fc9-b8cf-70a4eff73510-rbac\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.151740 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/49d8d406-58db-4226-806b-850e6e12b04f-tenants\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.154344 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhttz\" (UniqueName: \"kubernetes.io/projected/81042cee-7dc6-4fc9-b8cf-70a4eff73510-kube-api-access-fhttz\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.165355 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5qdf\" (UniqueName: \"kubernetes.io/projected/49d8d406-58db-4226-806b-850e6e12b04f-kube-api-access-l5qdf\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.199065 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/49d8d406-58db-4226-806b-850e6e12b04f-tls-secret\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.199157 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/81042cee-7dc6-4fc9-b8cf-70a4eff73510-tenants\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.199412 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/81042cee-7dc6-4fc9-b8cf-70a4eff73510-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-676969bccf-mxr7n\" (UID: \"81042cee-7dc6-4fc9-b8cf-70a4eff73510\") " pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.203396 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/49d8d406-58db-4226-806b-850e6e12b04f-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-676969bccf-hqhwj\" (UID: \"49d8d406-58db-4226-806b-850e6e12b04f\") " pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.457718 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.458385 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.554872 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.558111 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.590058 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.591276 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.612525 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.614394 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.617113 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.617258 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.625123 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.634114 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.671802 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.671856 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0ecc927f-14aa-4891-8411-7b98dfeb5c01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0ecc927f-14aa-4891-8411-7b98dfeb5c01\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.671882 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.671909 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.671933 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-config\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.671950 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cc1f76e4-d4f2-4ef3-a4c3-0e6186a97b6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cc1f76e4-d4f2-4ef3-a4c3-0e6186a97b6c\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.671999 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.672023 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s72h2\" (UniqueName: \"kubernetes.io/projected/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-kube-api-access-s72h2\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.723763 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.724645 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.727272 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.727410 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.743379 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773044 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0ecc927f-14aa-4891-8411-7b98dfeb5c01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0ecc927f-14aa-4891-8411-7b98dfeb5c01\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773464 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773492 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773522 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773542 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773559 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773577 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773600 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-config\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773633 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cc1f76e4-d4f2-4ef3-a4c3-0e6186a97b6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cc1f76e4-d4f2-4ef3-a4c3-0e6186a97b6c\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773653 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773674 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773691 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kxvr\" (UniqueName: \"kubernetes.io/projected/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-kube-api-access-5kxvr\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773724 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be42aebb-ed72-4e33-8e02-d8baace9c46d-config\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773750 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5160f111-f19a-497c-be33-3146dbfcbc81\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5160f111-f19a-497c-be33-3146dbfcbc81\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773769 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773785 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-config\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773807 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773825 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s72h2\" (UniqueName: \"kubernetes.io/projected/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-kube-api-access-s72h2\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773847 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a5eacc51-d2a2-4c2d-a3a4-e2eaabc4325c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a5eacc51-d2a2-4c2d-a3a4-e2eaabc4325c\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773864 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773884 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-878qd\" (UniqueName: \"kubernetes.io/projected/be42aebb-ed72-4e33-8e02-d8baace9c46d-kube-api-access-878qd\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.773903 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.775653 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.776320 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-config\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.779035 4954 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.779352 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cc1f76e4-d4f2-4ef3-a4c3-0e6186a97b6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cc1f76e4-d4f2-4ef3-a4c3-0e6186a97b6c\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e0179833483f9dedbf9db0a80f7b196c5532fba633dd1716995bb5335f812854/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.780149 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.780352 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.785032 4954 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.785190 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0ecc927f-14aa-4891-8411-7b98dfeb5c01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0ecc927f-14aa-4891-8411-7b98dfeb5c01\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5b1d1c704ba7b86361d17714265aee539c2a0fe9b5db99d34dd59598d2ac30c2/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.791192 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s72h2\" (UniqueName: \"kubernetes.io/projected/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-kube-api-access-s72h2\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.796109 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/392d7392-3bc6-42f1-91b5-56f00fc1d1c4-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.847621 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cc1f76e4-d4f2-4ef3-a4c3-0e6186a97b6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cc1f76e4-d4f2-4ef3-a4c3-0e6186a97b6c\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.888917 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.889084 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a5eacc51-d2a2-4c2d-a3a4-e2eaabc4325c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a5eacc51-d2a2-4c2d-a3a4-e2eaabc4325c\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.889126 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.889452 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-878qd\" (UniqueName: \"kubernetes.io/projected/be42aebb-ed72-4e33-8e02-d8baace9c46d-kube-api-access-878qd\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.890018 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.890158 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.891143 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.891764 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.893566 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.898150 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.899072 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.899107 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kxvr\" (UniqueName: \"kubernetes.io/projected/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-kube-api-access-5kxvr\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.899204 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be42aebb-ed72-4e33-8e02-d8baace9c46d-config\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.899260 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5160f111-f19a-497c-be33-3146dbfcbc81\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5160f111-f19a-497c-be33-3146dbfcbc81\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.899302 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-config\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.900968 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.920139 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-878qd\" (UniqueName: \"kubernetes.io/projected/be42aebb-ed72-4e33-8e02-d8baace9c46d-kube-api-access-878qd\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.937008 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.937451 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.937714 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.945771 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be42aebb-ed72-4e33-8e02-d8baace9c46d-config\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.946944 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.948194 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.955299 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-config\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.955972 4954 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.966875 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5160f111-f19a-497c-be33-3146dbfcbc81\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5160f111-f19a-497c-be33-3146dbfcbc81\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/359fb30e0b3e8fcf31924471e9561253976d4d0fc3cb3bdcb50763cfeb0ba746/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.967415 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/be42aebb-ed72-4e33-8e02-d8baace9c46d-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.968108 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0ecc927f-14aa-4891-8411-7b98dfeb5c01\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0ecc927f-14aa-4891-8411-7b98dfeb5c01\") pod \"logging-loki-ingester-0\" (UID: \"392d7392-3bc6-42f1-91b5-56f00fc1d1c4\") " pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.969653 4954 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.969683 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a5eacc51-d2a2-4c2d-a3a4-e2eaabc4325c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a5eacc51-d2a2-4c2d-a3a4-e2eaabc4325c\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fe093001ce721becf5baa8455f3347b58b4b6cc48df879509310ba8fd3af067c/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.972871 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr"] Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.975039 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kxvr\" (UniqueName: \"kubernetes.io/projected/7fd1add7-4d51-4474-a1c4-b95df89c1bb1-kube-api-access-5kxvr\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.989226 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:21 crc kubenswrapper[4954]: I1209 17:10:21.998565 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5160f111-f19a-497c-be33-3146dbfcbc81\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5160f111-f19a-497c-be33-3146dbfcbc81\") pod \"logging-loki-compactor-0\" (UID: \"7fd1add7-4d51-4474-a1c4-b95df89c1bb1\") " pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.017129 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a5eacc51-d2a2-4c2d-a3a4-e2eaabc4325c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a5eacc51-d2a2-4c2d-a3a4-e2eaabc4325c\") pod \"logging-loki-index-gateway-0\" (UID: \"be42aebb-ed72-4e33-8e02-d8baace9c46d\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.057422 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.087158 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-676969bccf-mxr7n"] Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.091998 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-stgnv"] Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.159965 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd"] Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.160321 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" event={"ID":"8ca5d9e5-51f2-4a7a-86e4-a862190a0828","Type":"ContainerStarted","Data":"3b09dcc78495c5a361257cafe93dcb4a9dc0812e7c4a215216c7609080b489eb"} Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.185682 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-676969bccf-hqhwj"] Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.299256 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.481653 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 09 17:10:22 crc kubenswrapper[4954]: W1209 17:10:22.493098 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe42aebb_ed72_4e33_8e02_d8baace9c46d.slice/crio-ae14d35ece6029029d7da372dae663b4f6c548ba65ebd42da20098017b021403 WatchSource:0}: Error finding container ae14d35ece6029029d7da372dae663b4f6c548ba65ebd42da20098017b021403: Status 404 returned error can't find the container with id ae14d35ece6029029d7da372dae663b4f6c548ba65ebd42da20098017b021403 Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.557220 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 09 17:10:22 crc kubenswrapper[4954]: W1209 17:10:22.560632 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fd1add7_4d51_4474_a1c4_b95df89c1bb1.slice/crio-d602dff1d77d42c24a8820a32cccfb6d0431df350ff0688dc04fd0d43f21fe15 WatchSource:0}: Error finding container d602dff1d77d42c24a8820a32cccfb6d0431df350ff0688dc04fd0d43f21fe15: Status 404 returned error can't find the container with id d602dff1d77d42c24a8820a32cccfb6d0431df350ff0688dc04fd0d43f21fe15 Dec 09 17:10:22 crc kubenswrapper[4954]: I1209 17:10:22.619328 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 09 17:10:23 crc kubenswrapper[4954]: I1209 17:10:23.137817 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"be42aebb-ed72-4e33-8e02-d8baace9c46d","Type":"ContainerStarted","Data":"ae14d35ece6029029d7da372dae663b4f6c548ba65ebd42da20098017b021403"} Dec 09 17:10:23 crc kubenswrapper[4954]: I1209 17:10:23.138672 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"392d7392-3bc6-42f1-91b5-56f00fc1d1c4","Type":"ContainerStarted","Data":"33994f9c96c28cdc3a3930c2b9e5ca2d9b6df669e566f19129dc694d86711a68"} Dec 09 17:10:23 crc kubenswrapper[4954]: I1209 17:10:23.139684 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" event={"ID":"f7dd2098-b496-49e7-9cdb-f87431dc135f","Type":"ContainerStarted","Data":"6b167e141e676581f4fb18d23f2e34bebbb12425468f1930b062c457be9afc98"} Dec 09 17:10:23 crc kubenswrapper[4954]: I1209 17:10:23.140818 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" event={"ID":"73361aee-2741-4328-b840-bcc3619f8cc0","Type":"ContainerStarted","Data":"c1e9f61bc93aa5fbd83d90ee866b6bbaef4c0c24562f0067816261d524b01c7a"} Dec 09 17:10:23 crc kubenswrapper[4954]: I1209 17:10:23.141982 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"7fd1add7-4d51-4474-a1c4-b95df89c1bb1","Type":"ContainerStarted","Data":"d602dff1d77d42c24a8820a32cccfb6d0431df350ff0688dc04fd0d43f21fe15"} Dec 09 17:10:23 crc kubenswrapper[4954]: I1209 17:10:23.143342 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" event={"ID":"49d8d406-58db-4226-806b-850e6e12b04f","Type":"ContainerStarted","Data":"1918d19724a9fdc0b17a2de55cf29279d3f6e6863719b4f74c35703ec09e2f44"} Dec 09 17:10:23 crc kubenswrapper[4954]: I1209 17:10:23.144484 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" event={"ID":"81042cee-7dc6-4fc9-b8cf-70a4eff73510","Type":"ContainerStarted","Data":"2b5a9447187d89ebbb10a4a2a5781368ed682bcecf7c83cb36314ed312038a01"} Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.325374 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" event={"ID":"8ca5d9e5-51f2-4a7a-86e4-a862190a0828","Type":"ContainerStarted","Data":"c927be149e57f1772f250b589c85bae96b486fea04a4a3676f5f74a7eaec018d"} Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.326043 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.326621 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"7fd1add7-4d51-4474-a1c4-b95df89c1bb1","Type":"ContainerStarted","Data":"bfd853ff01393ea8e647b187b97df74f22167b4bfaf634c22c84c98983cc50f8"} Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.327424 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.328508 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" event={"ID":"49d8d406-58db-4226-806b-850e6e12b04f","Type":"ContainerStarted","Data":"9abcf1b6c4bb802d9744d0863f6aa573300e9f94899a781d3dbf0dd439c3b4e6"} Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.329867 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" event={"ID":"81042cee-7dc6-4fc9-b8cf-70a4eff73510","Type":"ContainerStarted","Data":"b099ee5f5acfc9167ca9962299cded11bc345bef0331c7fcd61159322a99147f"} Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.330800 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"be42aebb-ed72-4e33-8e02-d8baace9c46d","Type":"ContainerStarted","Data":"cba13bd0ab1f68fff9a0e24a1e7d97ae1b94fbb021eb6fd4f28f8a9690646b82"} Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.331286 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.332280 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"392d7392-3bc6-42f1-91b5-56f00fc1d1c4","Type":"ContainerStarted","Data":"176fc44ffe2195304380aa4dd9e21054981586ee8f056dd64c6d72091fce6623"} Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.332697 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.333812 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" event={"ID":"f7dd2098-b496-49e7-9cdb-f87431dc135f","Type":"ContainerStarted","Data":"6ed66d1397a14dc4fc5fa277ff7bd2782da91c214bfbd1ba4272c93ed392b3ad"} Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.334194 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.335154 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" event={"ID":"73361aee-2741-4328-b840-bcc3619f8cc0","Type":"ContainerStarted","Data":"9b7206dce9dd6c4b0412d6f5385a2e18893ad278e42f61088c91f8044df20ae9"} Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.335632 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.364398 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" podStartSLOduration=2.533500823 podStartE2EDuration="10.364368391s" podCreationTimestamp="2025-12-09 17:10:20 +0000 UTC" firstStartedPulling="2025-12-09 17:10:21.979221366 +0000 UTC m=+818.367395186" lastFinishedPulling="2025-12-09 17:10:29.810088934 +0000 UTC m=+826.198262754" observedRunningTime="2025-12-09 17:10:30.346506542 +0000 UTC m=+826.734680362" watchObservedRunningTime="2025-12-09 17:10:30.364368391 +0000 UTC m=+826.752542211" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.413101 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" podStartSLOduration=2.78871527 podStartE2EDuration="10.413075965s" podCreationTimestamp="2025-12-09 17:10:20 +0000 UTC" firstStartedPulling="2025-12-09 17:10:22.177549903 +0000 UTC m=+818.565723733" lastFinishedPulling="2025-12-09 17:10:29.801910608 +0000 UTC m=+826.190084428" observedRunningTime="2025-12-09 17:10:30.392188482 +0000 UTC m=+826.780362302" watchObservedRunningTime="2025-12-09 17:10:30.413075965 +0000 UTC m=+826.801249785" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.418357 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.22900256 podStartE2EDuration="10.4183427s" podCreationTimestamp="2025-12-09 17:10:20 +0000 UTC" firstStartedPulling="2025-12-09 17:10:22.498985713 +0000 UTC m=+818.887159523" lastFinishedPulling="2025-12-09 17:10:29.688325843 +0000 UTC m=+826.076499663" observedRunningTime="2025-12-09 17:10:30.411329441 +0000 UTC m=+826.799503261" watchObservedRunningTime="2025-12-09 17:10:30.4183427 +0000 UTC m=+826.806516510" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.466344 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" podStartSLOduration=2.853566459 podStartE2EDuration="10.466318361s" podCreationTimestamp="2025-12-09 17:10:20 +0000 UTC" firstStartedPulling="2025-12-09 17:10:22.175046225 +0000 UTC m=+818.563220045" lastFinishedPulling="2025-12-09 17:10:29.787798127 +0000 UTC m=+826.175971947" observedRunningTime="2025-12-09 17:10:30.463450772 +0000 UTC m=+826.851624592" watchObservedRunningTime="2025-12-09 17:10:30.466318361 +0000 UTC m=+826.854492201" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.496469 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.271287472 podStartE2EDuration="10.496448395s" podCreationTimestamp="2025-12-09 17:10:20 +0000 UTC" firstStartedPulling="2025-12-09 17:10:22.562982025 +0000 UTC m=+818.951155845" lastFinishedPulling="2025-12-09 17:10:29.788142948 +0000 UTC m=+826.176316768" observedRunningTime="2025-12-09 17:10:30.493889654 +0000 UTC m=+826.882063474" watchObservedRunningTime="2025-12-09 17:10:30.496448395 +0000 UTC m=+826.884622215" Dec 09 17:10:30 crc kubenswrapper[4954]: I1209 17:10:30.522748 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=3.341851061 podStartE2EDuration="10.522726397s" podCreationTimestamp="2025-12-09 17:10:20 +0000 UTC" firstStartedPulling="2025-12-09 17:10:22.631886452 +0000 UTC m=+819.020060272" lastFinishedPulling="2025-12-09 17:10:29.812761788 +0000 UTC m=+826.200935608" observedRunningTime="2025-12-09 17:10:30.522449838 +0000 UTC m=+826.910623648" watchObservedRunningTime="2025-12-09 17:10:30.522726397 +0000 UTC m=+826.910900227" Dec 09 17:10:35 crc kubenswrapper[4954]: I1209 17:10:35.375016 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" event={"ID":"49d8d406-58db-4226-806b-850e6e12b04f","Type":"ContainerStarted","Data":"e0cd3638c0e0aca8322a7541401823b0a19851063ee387853d93ef105e8f7d2b"} Dec 09 17:10:35 crc kubenswrapper[4954]: I1209 17:10:35.375615 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:35 crc kubenswrapper[4954]: I1209 17:10:35.377910 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" event={"ID":"81042cee-7dc6-4fc9-b8cf-70a4eff73510","Type":"ContainerStarted","Data":"baa49ea48b0c190b2fe037296b9f57e314e039261040a1de1e361ba181e6e642"} Dec 09 17:10:35 crc kubenswrapper[4954]: I1209 17:10:35.378748 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:35 crc kubenswrapper[4954]: I1209 17:10:35.378788 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:35 crc kubenswrapper[4954]: I1209 17:10:35.387223 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:35 crc kubenswrapper[4954]: I1209 17:10:35.388794 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:35 crc kubenswrapper[4954]: I1209 17:10:35.396030 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" Dec 09 17:10:35 crc kubenswrapper[4954]: I1209 17:10:35.403685 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" podStartSLOduration=4.04184659 podStartE2EDuration="15.403659283s" podCreationTimestamp="2025-12-09 17:10:20 +0000 UTC" firstStartedPulling="2025-12-09 17:10:22.253904783 +0000 UTC m=+818.642078603" lastFinishedPulling="2025-12-09 17:10:33.615717476 +0000 UTC m=+830.003891296" observedRunningTime="2025-12-09 17:10:35.398348707 +0000 UTC m=+831.786522527" watchObservedRunningTime="2025-12-09 17:10:35.403659283 +0000 UTC m=+831.791833103" Dec 09 17:10:35 crc kubenswrapper[4954]: I1209 17:10:35.429521 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-676969bccf-mxr7n" podStartSLOduration=3.962837066 podStartE2EDuration="15.42948105s" podCreationTimestamp="2025-12-09 17:10:20 +0000 UTC" firstStartedPulling="2025-12-09 17:10:22.154534253 +0000 UTC m=+818.542708073" lastFinishedPulling="2025-12-09 17:10:33.621178237 +0000 UTC m=+830.009352057" observedRunningTime="2025-12-09 17:10:35.420410717 +0000 UTC m=+831.808584537" watchObservedRunningTime="2025-12-09 17:10:35.42948105 +0000 UTC m=+831.817654880" Dec 09 17:10:36 crc kubenswrapper[4954]: I1209 17:10:36.384460 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:36 crc kubenswrapper[4954]: I1209 17:10:36.397732 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-676969bccf-hqhwj" Dec 09 17:10:43 crc kubenswrapper[4954]: I1209 17:10:43.755742 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:10:43 crc kubenswrapper[4954]: I1209 17:10:43.757752 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:10:43 crc kubenswrapper[4954]: I1209 17:10:43.757841 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:10:43 crc kubenswrapper[4954]: I1209 17:10:43.758624 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ea6d5b701a579f932d38d9e389879e05d1afbb8ef2ed598058ec1bc38f02d1f5"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:10:43 crc kubenswrapper[4954]: I1209 17:10:43.758723 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://ea6d5b701a579f932d38d9e389879e05d1afbb8ef2ed598058ec1bc38f02d1f5" gracePeriod=600 Dec 09 17:10:44 crc kubenswrapper[4954]: I1209 17:10:44.455172 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="ea6d5b701a579f932d38d9e389879e05d1afbb8ef2ed598058ec1bc38f02d1f5" exitCode=0 Dec 09 17:10:44 crc kubenswrapper[4954]: I1209 17:10:44.455271 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"ea6d5b701a579f932d38d9e389879e05d1afbb8ef2ed598058ec1bc38f02d1f5"} Dec 09 17:10:44 crc kubenswrapper[4954]: I1209 17:10:44.455716 4954 scope.go:117] "RemoveContainer" containerID="f40e602b695fb058005b96319e15b3099ce90e1eb551fb5c6fef94df1b5d64c9" Dec 09 17:10:46 crc kubenswrapper[4954]: I1209 17:10:46.477510 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"809e10874828ead4756b40cdd641829a7300ce5c7ed7fd1b00389ceadbd9262b"} Dec 09 17:10:50 crc kubenswrapper[4954]: I1209 17:10:50.694106 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-hgwvr" Dec 09 17:10:50 crc kubenswrapper[4954]: I1209 17:10:50.849170 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-stgnv" Dec 09 17:10:51 crc kubenswrapper[4954]: I1209 17:10:51.014326 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-gzndd" Dec 09 17:10:51 crc kubenswrapper[4954]: I1209 17:10:51.996189 4954 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 09 17:10:51 crc kubenswrapper[4954]: I1209 17:10:51.996256 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="392d7392-3bc6-42f1-91b5-56f00fc1d1c4" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 17:10:52 crc kubenswrapper[4954]: I1209 17:10:52.065529 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 09 17:10:52 crc kubenswrapper[4954]: I1209 17:10:52.307264 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 09 17:11:01 crc kubenswrapper[4954]: I1209 17:11:01.996756 4954 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 09 17:11:01 crc kubenswrapper[4954]: I1209 17:11:01.997740 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="392d7392-3bc6-42f1-91b5-56f00fc1d1c4" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 17:11:11 crc kubenswrapper[4954]: I1209 17:11:11.997052 4954 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 09 17:11:11 crc kubenswrapper[4954]: I1209 17:11:11.997797 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="392d7392-3bc6-42f1-91b5-56f00fc1d1c4" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 17:11:21 crc kubenswrapper[4954]: I1209 17:11:21.994969 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 09 17:11:40 crc kubenswrapper[4954]: I1209 17:11:40.941233 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-qvzlk"] Dec 09 17:11:40 crc kubenswrapper[4954]: I1209 17:11:40.942742 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-qvzlk" Dec 09 17:11:40 crc kubenswrapper[4954]: I1209 17:11:40.945946 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-nbdck" Dec 09 17:11:40 crc kubenswrapper[4954]: I1209 17:11:40.946319 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 09 17:11:40 crc kubenswrapper[4954]: I1209 17:11:40.948473 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 09 17:11:40 crc kubenswrapper[4954]: I1209 17:11:40.949119 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 09 17:11:40 crc kubenswrapper[4954]: I1209 17:11:40.950918 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 09 17:11:40 crc kubenswrapper[4954]: I1209 17:11:40.955723 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 09 17:11:40 crc kubenswrapper[4954]: I1209 17:11:40.977832 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-qvzlk"] Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.037499 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/744189b4-da3e-4dab-a470-e7a6dc068066-datadir\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.037634 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-metrics\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.037767 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config-openshift-service-cacrt\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.037789 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97qbx\" (UniqueName: \"kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-kube-api-access-97qbx\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.037852 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-entrypoint\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.038092 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-trusted-ca\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.038233 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/744189b4-da3e-4dab-a470-e7a6dc068066-tmp\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.038302 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-sa-token\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.038374 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.038412 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-syslog-receiver\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.038456 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-token\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.048470 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-qvzlk"] Dec 09 17:11:41 crc kubenswrapper[4954]: E1209 17:11:41.049057 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-97qbx metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-qvzlk" podUID="744189b4-da3e-4dab-a470-e7a6dc068066" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.140767 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/744189b4-da3e-4dab-a470-e7a6dc068066-datadir\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.140852 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-metrics\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.140877 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config-openshift-service-cacrt\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.140894 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97qbx\" (UniqueName: \"kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-kube-api-access-97qbx\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.140948 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-entrypoint\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.141004 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-trusted-ca\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.141025 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/744189b4-da3e-4dab-a470-e7a6dc068066-tmp\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.141042 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-sa-token\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.141080 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.141098 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-syslog-receiver\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.141117 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-token\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: E1209 17:11:41.141742 4954 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Dec 09 17:11:41 crc kubenswrapper[4954]: E1209 17:11:41.141829 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-syslog-receiver podName:744189b4-da3e-4dab-a470-e7a6dc068066 nodeName:}" failed. No retries permitted until 2025-12-09 17:11:41.641805907 +0000 UTC m=+898.029979727 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-syslog-receiver") pod "collector-qvzlk" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066") : secret "collector-syslog-receiver" not found Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.141739 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/744189b4-da3e-4dab-a470-e7a6dc068066-datadir\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.142654 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-entrypoint\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.142747 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config-openshift-service-cacrt\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.142967 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.143544 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-trusted-ca\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.151199 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-metrics\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.151826 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-token\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.152445 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/744189b4-da3e-4dab-a470-e7a6dc068066-tmp\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.166926 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-sa-token\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.179738 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97qbx\" (UniqueName: \"kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-kube-api-access-97qbx\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.648709 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-syslog-receiver\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.655542 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-syslog-receiver\") pod \"collector-qvzlk\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.863938 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.876918 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-qvzlk" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.953496 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config-openshift-service-cacrt\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.953786 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-entrypoint\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.953896 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-token\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.953986 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-sa-token\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.954062 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.954076 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/744189b4-da3e-4dab-a470-e7a6dc068066-tmp\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.954165 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-metrics\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.954254 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97qbx\" (UniqueName: \"kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-kube-api-access-97qbx\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.954306 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-syslog-receiver\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.954332 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-trusted-ca\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.954407 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/744189b4-da3e-4dab-a470-e7a6dc068066-datadir\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.954439 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config\") pod \"744189b4-da3e-4dab-a470-e7a6dc068066\" (UID: \"744189b4-da3e-4dab-a470-e7a6dc068066\") " Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.954794 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.955156 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config" (OuterVolumeSpecName: "config") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.955302 4954 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.955324 4954 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-entrypoint\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.955817 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/744189b4-da3e-4dab-a470-e7a6dc068066-datadir" (OuterVolumeSpecName: "datadir") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.956318 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.958851 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-metrics" (OuterVolumeSpecName: "metrics") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.959555 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/744189b4-da3e-4dab-a470-e7a6dc068066-tmp" (OuterVolumeSpecName: "tmp") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.959564 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-kube-api-access-97qbx" (OuterVolumeSpecName: "kube-api-access-97qbx") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "kube-api-access-97qbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.960375 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-sa-token" (OuterVolumeSpecName: "sa-token") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.960532 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:11:41 crc kubenswrapper[4954]: I1209 17:11:41.965738 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-token" (OuterVolumeSpecName: "collector-token") pod "744189b4-da3e-4dab-a470-e7a6dc068066" (UID: "744189b4-da3e-4dab-a470-e7a6dc068066"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.056952 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97qbx\" (UniqueName: \"kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-kube-api-access-97qbx\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.056988 4954 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.056998 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.057027 4954 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/744189b4-da3e-4dab-a470-e7a6dc068066-datadir\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.057036 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/744189b4-da3e-4dab-a470-e7a6dc068066-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.057045 4954 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-collector-token\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.057053 4954 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/744189b4-da3e-4dab-a470-e7a6dc068066-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.057062 4954 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/744189b4-da3e-4dab-a470-e7a6dc068066-tmp\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.057070 4954 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/744189b4-da3e-4dab-a470-e7a6dc068066-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.869136 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-qvzlk" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.925677 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-qvzlk"] Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.930705 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-qvzlk"] Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.935879 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-87lsn"] Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.936889 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-87lsn" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.943386 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.943424 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-87lsn"] Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.945944 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-nbdck" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.946012 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.945944 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.946243 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 09 17:11:42 crc kubenswrapper[4954]: I1209 17:11:42.953336 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.073217 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7302a86b-a768-46f5-af6b-e3429564341d-metrics\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.073870 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7302a86b-a768-46f5-af6b-e3429564341d-collector-syslog-receiver\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.073901 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-entrypoint\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.073935 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4dll\" (UniqueName: \"kubernetes.io/projected/7302a86b-a768-46f5-af6b-e3429564341d-kube-api-access-q4dll\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.073967 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-trusted-ca\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.074105 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-config\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.074255 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7302a86b-a768-46f5-af6b-e3429564341d-tmp\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.074335 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7302a86b-a768-46f5-af6b-e3429564341d-collector-token\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.074355 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-config-openshift-service-cacrt\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.074379 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7302a86b-a768-46f5-af6b-e3429564341d-sa-token\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.074402 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7302a86b-a768-46f5-af6b-e3429564341d-datadir\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.176500 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7302a86b-a768-46f5-af6b-e3429564341d-collector-syslog-receiver\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.176884 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-entrypoint\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.177012 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4dll\" (UniqueName: \"kubernetes.io/projected/7302a86b-a768-46f5-af6b-e3429564341d-kube-api-access-q4dll\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.177123 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-trusted-ca\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.177224 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-config\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.177342 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7302a86b-a768-46f5-af6b-e3429564341d-tmp\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.177466 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7302a86b-a768-46f5-af6b-e3429564341d-collector-token\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.177608 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-config-openshift-service-cacrt\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.177720 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7302a86b-a768-46f5-af6b-e3429564341d-sa-token\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.177836 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7302a86b-a768-46f5-af6b-e3429564341d-datadir\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.177957 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7302a86b-a768-46f5-af6b-e3429564341d-datadir\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.178077 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7302a86b-a768-46f5-af6b-e3429564341d-metrics\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.178442 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-entrypoint\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.178620 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-config-openshift-service-cacrt\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.178784 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-trusted-ca\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.178939 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7302a86b-a768-46f5-af6b-e3429564341d-config\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.180949 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7302a86b-a768-46f5-af6b-e3429564341d-metrics\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.183571 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7302a86b-a768-46f5-af6b-e3429564341d-tmp\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.185495 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7302a86b-a768-46f5-af6b-e3429564341d-collector-token\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.186909 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7302a86b-a768-46f5-af6b-e3429564341d-collector-syslog-receiver\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.193896 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4dll\" (UniqueName: \"kubernetes.io/projected/7302a86b-a768-46f5-af6b-e3429564341d-kube-api-access-q4dll\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.194842 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7302a86b-a768-46f5-af6b-e3429564341d-sa-token\") pod \"collector-87lsn\" (UID: \"7302a86b-a768-46f5-af6b-e3429564341d\") " pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.258104 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-87lsn" Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.675747 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-87lsn"] Dec 09 17:11:43 crc kubenswrapper[4954]: I1209 17:11:43.880744 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-87lsn" event={"ID":"7302a86b-a768-46f5-af6b-e3429564341d","Type":"ContainerStarted","Data":"2b3a6f4423367632cf87c32af230a96c46965bb5515e2068ea19439c3dfca936"} Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.130011 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="744189b4-da3e-4dab-a470-e7a6dc068066" path="/var/lib/kubelet/pods/744189b4-da3e-4dab-a470-e7a6dc068066/volumes" Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.751793 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hxkhn"] Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.753421 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.765493 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxkhn"] Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.802566 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-catalog-content\") pod \"redhat-marketplace-hxkhn\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.802658 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-utilities\") pod \"redhat-marketplace-hxkhn\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.802732 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs4xz\" (UniqueName: \"kubernetes.io/projected/d0d5f9e7-dc11-4753-9045-c38b82df963b-kube-api-access-rs4xz\") pod \"redhat-marketplace-hxkhn\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.903888 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-utilities\") pod \"redhat-marketplace-hxkhn\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.903997 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs4xz\" (UniqueName: \"kubernetes.io/projected/d0d5f9e7-dc11-4753-9045-c38b82df963b-kube-api-access-rs4xz\") pod \"redhat-marketplace-hxkhn\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.904061 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-catalog-content\") pod \"redhat-marketplace-hxkhn\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.904500 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-utilities\") pod \"redhat-marketplace-hxkhn\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.904539 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-catalog-content\") pod \"redhat-marketplace-hxkhn\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:44 crc kubenswrapper[4954]: I1209 17:11:44.926987 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs4xz\" (UniqueName: \"kubernetes.io/projected/d0d5f9e7-dc11-4753-9045-c38b82df963b-kube-api-access-rs4xz\") pod \"redhat-marketplace-hxkhn\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:45 crc kubenswrapper[4954]: I1209 17:11:45.076642 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:45 crc kubenswrapper[4954]: I1209 17:11:45.295374 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxkhn"] Dec 09 17:11:45 crc kubenswrapper[4954]: I1209 17:11:45.895789 4954 generic.go:334] "Generic (PLEG): container finished" podID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerID="a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe" exitCode=0 Dec 09 17:11:45 crc kubenswrapper[4954]: I1209 17:11:45.895857 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxkhn" event={"ID":"d0d5f9e7-dc11-4753-9045-c38b82df963b","Type":"ContainerDied","Data":"a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe"} Dec 09 17:11:45 crc kubenswrapper[4954]: I1209 17:11:45.895896 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxkhn" event={"ID":"d0d5f9e7-dc11-4753-9045-c38b82df963b","Type":"ContainerStarted","Data":"f49545691f15c7a74fe5665f73a90ffd5c897016b27b362003a6fe129497fb64"} Dec 09 17:11:46 crc kubenswrapper[4954]: I1209 17:11:46.908863 4954 generic.go:334] "Generic (PLEG): container finished" podID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerID="7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48" exitCode=0 Dec 09 17:11:46 crc kubenswrapper[4954]: I1209 17:11:46.908937 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxkhn" event={"ID":"d0d5f9e7-dc11-4753-9045-c38b82df963b","Type":"ContainerDied","Data":"7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48"} Dec 09 17:11:48 crc kubenswrapper[4954]: I1209 17:11:47.917952 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxkhn" event={"ID":"d0d5f9e7-dc11-4753-9045-c38b82df963b","Type":"ContainerStarted","Data":"a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784"} Dec 09 17:11:48 crc kubenswrapper[4954]: I1209 17:11:47.945293 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hxkhn" podStartSLOduration=2.533759363 podStartE2EDuration="3.945267765s" podCreationTimestamp="2025-12-09 17:11:44 +0000 UTC" firstStartedPulling="2025-12-09 17:11:45.899491125 +0000 UTC m=+902.287664945" lastFinishedPulling="2025-12-09 17:11:47.310999507 +0000 UTC m=+903.699173347" observedRunningTime="2025-12-09 17:11:47.940228008 +0000 UTC m=+904.328401868" watchObservedRunningTime="2025-12-09 17:11:47.945267765 +0000 UTC m=+904.333441595" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.157966 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-89cw6"] Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.161582 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.182450 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-89cw6"] Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.212234 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-utilities\") pod \"community-operators-89cw6\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.212322 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-catalog-content\") pod \"community-operators-89cw6\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.212378 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzx9z\" (UniqueName: \"kubernetes.io/projected/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-kube-api-access-nzx9z\") pod \"community-operators-89cw6\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.313050 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-utilities\") pod \"community-operators-89cw6\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.313103 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-catalog-content\") pod \"community-operators-89cw6\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.313139 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzx9z\" (UniqueName: \"kubernetes.io/projected/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-kube-api-access-nzx9z\") pod \"community-operators-89cw6\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.313871 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-utilities\") pod \"community-operators-89cw6\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.314001 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-catalog-content\") pod \"community-operators-89cw6\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.347027 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzx9z\" (UniqueName: \"kubernetes.io/projected/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-kube-api-access-nzx9z\") pod \"community-operators-89cw6\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:50 crc kubenswrapper[4954]: I1209 17:11:50.488435 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:11:51 crc kubenswrapper[4954]: I1209 17:11:51.399679 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-89cw6"] Dec 09 17:11:51 crc kubenswrapper[4954]: I1209 17:11:51.956103 4954 generic.go:334] "Generic (PLEG): container finished" podID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerID="c2a1bff4736a10e68192e59f51324e1edcc4b33851f0ae5e7ccca4c4c92eb015" exitCode=0 Dec 09 17:11:51 crc kubenswrapper[4954]: I1209 17:11:51.956218 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89cw6" event={"ID":"eacf2231-6eae-4b3c-aaea-ee0d90c4a150","Type":"ContainerDied","Data":"c2a1bff4736a10e68192e59f51324e1edcc4b33851f0ae5e7ccca4c4c92eb015"} Dec 09 17:11:51 crc kubenswrapper[4954]: I1209 17:11:51.956444 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89cw6" event={"ID":"eacf2231-6eae-4b3c-aaea-ee0d90c4a150","Type":"ContainerStarted","Data":"7c5cf4a9c0bf1968e0082ccafa3b7a58517384b9630fdc411db5f6acee453972"} Dec 09 17:11:55 crc kubenswrapper[4954]: I1209 17:11:55.077475 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:55 crc kubenswrapper[4954]: I1209 17:11:55.077944 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:55 crc kubenswrapper[4954]: I1209 17:11:55.192581 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:56 crc kubenswrapper[4954]: I1209 17:11:56.036691 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:56 crc kubenswrapper[4954]: I1209 17:11:56.316158 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxkhn"] Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.000882 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hxkhn" podUID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerName="registry-server" containerID="cri-o://a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784" gracePeriod=2 Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.494654 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.848832 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-utilities\") pod \"d0d5f9e7-dc11-4753-9045-c38b82df963b\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.848954 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-catalog-content\") pod \"d0d5f9e7-dc11-4753-9045-c38b82df963b\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.849047 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs4xz\" (UniqueName: \"kubernetes.io/projected/d0d5f9e7-dc11-4753-9045-c38b82df963b-kube-api-access-rs4xz\") pod \"d0d5f9e7-dc11-4753-9045-c38b82df963b\" (UID: \"d0d5f9e7-dc11-4753-9045-c38b82df963b\") " Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.850100 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-utilities" (OuterVolumeSpecName: "utilities") pod "d0d5f9e7-dc11-4753-9045-c38b82df963b" (UID: "d0d5f9e7-dc11-4753-9045-c38b82df963b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.858984 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d5f9e7-dc11-4753-9045-c38b82df963b-kube-api-access-rs4xz" (OuterVolumeSpecName: "kube-api-access-rs4xz") pod "d0d5f9e7-dc11-4753-9045-c38b82df963b" (UID: "d0d5f9e7-dc11-4753-9045-c38b82df963b"). InnerVolumeSpecName "kube-api-access-rs4xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.870150 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0d5f9e7-dc11-4753-9045-c38b82df963b" (UID: "d0d5f9e7-dc11-4753-9045-c38b82df963b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.951901 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.951958 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0d5f9e7-dc11-4753-9045-c38b82df963b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:58 crc kubenswrapper[4954]: I1209 17:11:58.951973 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs4xz\" (UniqueName: \"kubernetes.io/projected/d0d5f9e7-dc11-4753-9045-c38b82df963b-kube-api-access-rs4xz\") on node \"crc\" DevicePath \"\"" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.013187 4954 generic.go:334] "Generic (PLEG): container finished" podID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerID="a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784" exitCode=0 Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.013313 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxkhn" event={"ID":"d0d5f9e7-dc11-4753-9045-c38b82df963b","Type":"ContainerDied","Data":"a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784"} Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.013363 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxkhn" event={"ID":"d0d5f9e7-dc11-4753-9045-c38b82df963b","Type":"ContainerDied","Data":"f49545691f15c7a74fe5665f73a90ffd5c897016b27b362003a6fe129497fb64"} Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.013396 4954 scope.go:117] "RemoveContainer" containerID="a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.013627 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxkhn" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.020021 4954 generic.go:334] "Generic (PLEG): container finished" podID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerID="07bd93f1f28866b091c0bd124fce07b4df961f2810726c9eece5724556ccd997" exitCode=0 Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.020096 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89cw6" event={"ID":"eacf2231-6eae-4b3c-aaea-ee0d90c4a150","Type":"ContainerDied","Data":"07bd93f1f28866b091c0bd124fce07b4df961f2810726c9eece5724556ccd997"} Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.022667 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-87lsn" event={"ID":"7302a86b-a768-46f5-af6b-e3429564341d","Type":"ContainerStarted","Data":"9ec53c083d303fe24010312066019171ae179accf91e2d87b0ffd57c990ac85e"} Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.041294 4954 scope.go:117] "RemoveContainer" containerID="7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.077594 4954 scope.go:117] "RemoveContainer" containerID="a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.077936 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-87lsn" podStartSLOduration=2.497194893 podStartE2EDuration="17.077808211s" podCreationTimestamp="2025-12-09 17:11:42 +0000 UTC" firstStartedPulling="2025-12-09 17:11:43.682890501 +0000 UTC m=+900.071064321" lastFinishedPulling="2025-12-09 17:11:58.263503819 +0000 UTC m=+914.651677639" observedRunningTime="2025-12-09 17:11:59.071372259 +0000 UTC m=+915.459546079" watchObservedRunningTime="2025-12-09 17:11:59.077808211 +0000 UTC m=+915.465982031" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.100226 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxkhn"] Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.110396 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxkhn"] Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.113879 4954 scope.go:117] "RemoveContainer" containerID="a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784" Dec 09 17:11:59 crc kubenswrapper[4954]: E1209 17:11:59.114667 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784\": container with ID starting with a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784 not found: ID does not exist" containerID="a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.114708 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784"} err="failed to get container status \"a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784\": rpc error: code = NotFound desc = could not find container \"a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784\": container with ID starting with a066df00c00d8a8e3bffe44b0c724c2e2a5fc9b592dec2e0ee10587df7f12784 not found: ID does not exist" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.114743 4954 scope.go:117] "RemoveContainer" containerID="7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48" Dec 09 17:11:59 crc kubenswrapper[4954]: E1209 17:11:59.115324 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48\": container with ID starting with 7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48 not found: ID does not exist" containerID="7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.115368 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48"} err="failed to get container status \"7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48\": rpc error: code = NotFound desc = could not find container \"7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48\": container with ID starting with 7a1bced75b519a90251d99a16e69b26718e5bbad1289cce3d7f05057fa131a48 not found: ID does not exist" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.115397 4954 scope.go:117] "RemoveContainer" containerID="a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe" Dec 09 17:11:59 crc kubenswrapper[4954]: E1209 17:11:59.115890 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe\": container with ID starting with a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe not found: ID does not exist" containerID="a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe" Dec 09 17:11:59 crc kubenswrapper[4954]: I1209 17:11:59.115957 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe"} err="failed to get container status \"a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe\": rpc error: code = NotFound desc = could not find container \"a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe\": container with ID starting with a513b4d68e89283bfa8882a3c95e85ca7a243bf7f261efb70d1360fcfafb4cbe not found: ID does not exist" Dec 09 17:12:00 crc kubenswrapper[4954]: I1209 17:12:00.032456 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89cw6" event={"ID":"eacf2231-6eae-4b3c-aaea-ee0d90c4a150","Type":"ContainerStarted","Data":"b1e17332acffcfd89f8db2a0c9f46aeb96232f2da01614b866f1ed4a97db9c21"} Dec 09 17:12:00 crc kubenswrapper[4954]: I1209 17:12:00.055826 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-89cw6" podStartSLOduration=7.412356136 podStartE2EDuration="10.055809331s" podCreationTimestamp="2025-12-09 17:11:50 +0000 UTC" firstStartedPulling="2025-12-09 17:11:56.959039257 +0000 UTC m=+913.347213077" lastFinishedPulling="2025-12-09 17:11:59.602492452 +0000 UTC m=+915.990666272" observedRunningTime="2025-12-09 17:12:00.051702833 +0000 UTC m=+916.439876673" watchObservedRunningTime="2025-12-09 17:12:00.055809331 +0000 UTC m=+916.443983151" Dec 09 17:12:00 crc kubenswrapper[4954]: I1209 17:12:00.129061 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0d5f9e7-dc11-4753-9045-c38b82df963b" path="/var/lib/kubelet/pods/d0d5f9e7-dc11-4753-9045-c38b82df963b/volumes" Dec 09 17:12:00 crc kubenswrapper[4954]: I1209 17:12:00.489292 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:12:00 crc kubenswrapper[4954]: I1209 17:12:00.489375 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:12:01 crc kubenswrapper[4954]: I1209 17:12:01.545030 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-89cw6" podUID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerName="registry-server" probeResult="failure" output=< Dec 09 17:12:01 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 17:12:01 crc kubenswrapper[4954]: > Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.324038 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-td9wp"] Dec 09 17:12:02 crc kubenswrapper[4954]: E1209 17:12:02.324458 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerName="extract-utilities" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.324489 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerName="extract-utilities" Dec 09 17:12:02 crc kubenswrapper[4954]: E1209 17:12:02.324502 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerName="extract-content" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.324510 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerName="extract-content" Dec 09 17:12:02 crc kubenswrapper[4954]: E1209 17:12:02.324528 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerName="registry-server" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.324534 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerName="registry-server" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.324693 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d5f9e7-dc11-4753-9045-c38b82df963b" containerName="registry-server" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.326134 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.341972 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-td9wp"] Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.415509 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-utilities\") pod \"certified-operators-td9wp\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.415680 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-catalog-content\") pod \"certified-operators-td9wp\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.415727 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgctk\" (UniqueName: \"kubernetes.io/projected/e37a9f7e-c431-4ee3-8286-4716a36f1f63-kube-api-access-mgctk\") pod \"certified-operators-td9wp\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.517183 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-catalog-content\") pod \"certified-operators-td9wp\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.517251 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgctk\" (UniqueName: \"kubernetes.io/projected/e37a9f7e-c431-4ee3-8286-4716a36f1f63-kube-api-access-mgctk\") pod \"certified-operators-td9wp\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.517309 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-utilities\") pod \"certified-operators-td9wp\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.517920 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-utilities\") pod \"certified-operators-td9wp\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.518054 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-catalog-content\") pod \"certified-operators-td9wp\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.541329 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgctk\" (UniqueName: \"kubernetes.io/projected/e37a9f7e-c431-4ee3-8286-4716a36f1f63-kube-api-access-mgctk\") pod \"certified-operators-td9wp\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:02 crc kubenswrapper[4954]: I1209 17:12:02.652183 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:03 crc kubenswrapper[4954]: I1209 17:12:03.247530 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-td9wp"] Dec 09 17:12:04 crc kubenswrapper[4954]: I1209 17:12:04.066110 4954 generic.go:334] "Generic (PLEG): container finished" podID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerID="d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18" exitCode=0 Dec 09 17:12:04 crc kubenswrapper[4954]: I1209 17:12:04.066211 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9wp" event={"ID":"e37a9f7e-c431-4ee3-8286-4716a36f1f63","Type":"ContainerDied","Data":"d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18"} Dec 09 17:12:04 crc kubenswrapper[4954]: I1209 17:12:04.068953 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9wp" event={"ID":"e37a9f7e-c431-4ee3-8286-4716a36f1f63","Type":"ContainerStarted","Data":"8360cb043f056f63f2d7c8446b40b5618a1458f14b8f389e3f0343b105e6096a"} Dec 09 17:12:07 crc kubenswrapper[4954]: I1209 17:12:07.093263 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9wp" event={"ID":"e37a9f7e-c431-4ee3-8286-4716a36f1f63","Type":"ContainerStarted","Data":"c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3"} Dec 09 17:12:08 crc kubenswrapper[4954]: I1209 17:12:08.104623 4954 generic.go:334] "Generic (PLEG): container finished" podID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerID="c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3" exitCode=0 Dec 09 17:12:08 crc kubenswrapper[4954]: I1209 17:12:08.104668 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9wp" event={"ID":"e37a9f7e-c431-4ee3-8286-4716a36f1f63","Type":"ContainerDied","Data":"c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3"} Dec 09 17:12:09 crc kubenswrapper[4954]: I1209 17:12:09.112058 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9wp" event={"ID":"e37a9f7e-c431-4ee3-8286-4716a36f1f63","Type":"ContainerStarted","Data":"dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee"} Dec 09 17:12:09 crc kubenswrapper[4954]: I1209 17:12:09.146449 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-td9wp" podStartSLOduration=2.613307556 podStartE2EDuration="7.146413877s" podCreationTimestamp="2025-12-09 17:12:02 +0000 UTC" firstStartedPulling="2025-12-09 17:12:04.068958809 +0000 UTC m=+920.457132629" lastFinishedPulling="2025-12-09 17:12:08.60206513 +0000 UTC m=+924.990238950" observedRunningTime="2025-12-09 17:12:09.131479029 +0000 UTC m=+925.519652859" watchObservedRunningTime="2025-12-09 17:12:09.146413877 +0000 UTC m=+925.534587697" Dec 09 17:12:10 crc kubenswrapper[4954]: I1209 17:12:10.550327 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:12:10 crc kubenswrapper[4954]: I1209 17:12:10.595400 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:12:10 crc kubenswrapper[4954]: I1209 17:12:10.911721 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-89cw6"] Dec 09 17:12:12 crc kubenswrapper[4954]: I1209 17:12:12.134137 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-89cw6" podUID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerName="registry-server" containerID="cri-o://b1e17332acffcfd89f8db2a0c9f46aeb96232f2da01614b866f1ed4a97db9c21" gracePeriod=2 Dec 09 17:12:12 crc kubenswrapper[4954]: I1209 17:12:12.653358 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:12 crc kubenswrapper[4954]: I1209 17:12:12.653938 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:12 crc kubenswrapper[4954]: I1209 17:12:12.698576 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.145683 4954 generic.go:334] "Generic (PLEG): container finished" podID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerID="b1e17332acffcfd89f8db2a0c9f46aeb96232f2da01614b866f1ed4a97db9c21" exitCode=0 Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.145766 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89cw6" event={"ID":"eacf2231-6eae-4b3c-aaea-ee0d90c4a150","Type":"ContainerDied","Data":"b1e17332acffcfd89f8db2a0c9f46aeb96232f2da01614b866f1ed4a97db9c21"} Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.187836 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.730186 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.813968 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.820229 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb848vsz"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.829277 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.835489 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210blxw5"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.837620 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-catalog-content\") pod \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.837751 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzx9z\" (UniqueName: \"kubernetes.io/projected/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-kube-api-access-nzx9z\") pod \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.837796 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-utilities\") pod \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\" (UID: \"eacf2231-6eae-4b3c-aaea-ee0d90c4a150\") " Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.839050 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-utilities" (OuterVolumeSpecName: "utilities") pod "eacf2231-6eae-4b3c-aaea-ee0d90c4a150" (UID: "eacf2231-6eae-4b3c-aaea-ee0d90c4a150"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.851058 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-kube-api-access-nzx9z" (OuterVolumeSpecName: "kube-api-access-nzx9z") pod "eacf2231-6eae-4b3c-aaea-ee0d90c4a150" (UID: "eacf2231-6eae-4b3c-aaea-ee0d90c4a150"). InnerVolumeSpecName "kube-api-access-nzx9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.861451 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.865945 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f6sl7n"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.872371 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-45kb2"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.872923 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-45kb2" podUID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerName="registry-server" containerID="cri-o://72df4cc2d6b4b84c5cb6dfc9d53e848945487680f3f16efd084b2850dffea7f9" gracePeriod=30 Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.883239 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-td9wp"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.890672 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g4cpr"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.891030 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g4cpr" podUID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerName="registry-server" containerID="cri-o://697f4a455885134b6769d2283d6a950e72e67e1cf3a4ddab3183dbeeabab9213" gracePeriod=30 Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.904752 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v98r2"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.905090 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" podUID="6ff181e0-303d-4c55-9fc5-8c80561714ca" containerName="marketplace-operator" containerID="cri-o://33195b1264b7a3fb8bf839bf88ae5b95c78bae9ae678023058d9a2aee9eab833" gracePeriod=30 Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.905170 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eacf2231-6eae-4b3c-aaea-ee0d90c4a150" (UID: "eacf2231-6eae-4b3c-aaea-ee0d90c4a150"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.910385 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7x66v"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.910872 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7x66v" podUID="4084f661-2955-43ba-a2d4-e6389830d100" containerName="registry-server" containerID="cri-o://35cdf07ce6733998fff22a56eb25f59a8ba9b435366fa01e6aac64db13999372" gracePeriod=30 Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.924677 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4svl9"] Dec 09 17:12:13 crc kubenswrapper[4954]: E1209 17:12:13.925171 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerName="extract-utilities" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.925189 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerName="extract-utilities" Dec 09 17:12:13 crc kubenswrapper[4954]: E1209 17:12:13.925209 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerName="registry-server" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.925216 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerName="registry-server" Dec 09 17:12:13 crc kubenswrapper[4954]: E1209 17:12:13.925228 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerName="extract-content" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.925235 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerName="extract-content" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.925415 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" containerName="registry-server" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.926321 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.929005 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8lch"] Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.929287 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q8lch" podUID="30f323d5-cbe3-4209-a708-a2e277eab190" containerName="registry-server" containerID="cri-o://9289442f940546cbcd543822d629ae49e411d9d8eb95f4dde17e9bebf5ff4156" gracePeriod=30 Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.939643 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2fd60f62-8fed-4f13-9da7-926b99b7d4aa-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4svl9\" (UID: \"2fd60f62-8fed-4f13-9da7-926b99b7d4aa\") " pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.939745 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2fd60f62-8fed-4f13-9da7-926b99b7d4aa-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4svl9\" (UID: \"2fd60f62-8fed-4f13-9da7-926b99b7d4aa\") " pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.939810 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh22v\" (UniqueName: \"kubernetes.io/projected/2fd60f62-8fed-4f13-9da7-926b99b7d4aa-kube-api-access-jh22v\") pod \"marketplace-operator-79b997595-4svl9\" (UID: \"2fd60f62-8fed-4f13-9da7-926b99b7d4aa\") " pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.939859 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.939875 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzx9z\" (UniqueName: \"kubernetes.io/projected/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-kube-api-access-nzx9z\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.939890 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eacf2231-6eae-4b3c-aaea-ee0d90c4a150-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:13 crc kubenswrapper[4954]: I1209 17:12:13.944978 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4svl9"] Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.042268 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2fd60f62-8fed-4f13-9da7-926b99b7d4aa-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4svl9\" (UID: \"2fd60f62-8fed-4f13-9da7-926b99b7d4aa\") " pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.042810 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh22v\" (UniqueName: \"kubernetes.io/projected/2fd60f62-8fed-4f13-9da7-926b99b7d4aa-kube-api-access-jh22v\") pod \"marketplace-operator-79b997595-4svl9\" (UID: \"2fd60f62-8fed-4f13-9da7-926b99b7d4aa\") " pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.042868 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2fd60f62-8fed-4f13-9da7-926b99b7d4aa-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4svl9\" (UID: \"2fd60f62-8fed-4f13-9da7-926b99b7d4aa\") " pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.043735 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2fd60f62-8fed-4f13-9da7-926b99b7d4aa-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4svl9\" (UID: \"2fd60f62-8fed-4f13-9da7-926b99b7d4aa\") " pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.047940 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2fd60f62-8fed-4f13-9da7-926b99b7d4aa-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4svl9\" (UID: \"2fd60f62-8fed-4f13-9da7-926b99b7d4aa\") " pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.065014 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh22v\" (UniqueName: \"kubernetes.io/projected/2fd60f62-8fed-4f13-9da7-926b99b7d4aa-kube-api-access-jh22v\") pod \"marketplace-operator-79b997595-4svl9\" (UID: \"2fd60f62-8fed-4f13-9da7-926b99b7d4aa\") " pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.142011 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39ee90e9-4a6f-4f98-9dab-da37a93221f2" path="/var/lib/kubelet/pods/39ee90e9-4a6f-4f98-9dab-da37a93221f2/volumes" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.142862 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="548caf7d-1620-4634-b87f-b7c6049386da" path="/var/lib/kubelet/pods/548caf7d-1620-4634-b87f-b7c6049386da/volumes" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.144250 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e59b591a-d66f-4c27-9151-fa15ff9f84cc" path="/var/lib/kubelet/pods/e59b591a-d66f-4c27-9151-fa15ff9f84cc/volumes" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.168781 4954 generic.go:334] "Generic (PLEG): container finished" podID="30f323d5-cbe3-4209-a708-a2e277eab190" containerID="9289442f940546cbcd543822d629ae49e411d9d8eb95f4dde17e9bebf5ff4156" exitCode=0 Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.168882 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8lch" event={"ID":"30f323d5-cbe3-4209-a708-a2e277eab190","Type":"ContainerDied","Data":"9289442f940546cbcd543822d629ae49e411d9d8eb95f4dde17e9bebf5ff4156"} Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.171559 4954 generic.go:334] "Generic (PLEG): container finished" podID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerID="697f4a455885134b6769d2283d6a950e72e67e1cf3a4ddab3183dbeeabab9213" exitCode=0 Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.171662 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4cpr" event={"ID":"da66ed1d-0022-4970-bc4c-aa6a62df0755","Type":"ContainerDied","Data":"697f4a455885134b6769d2283d6a950e72e67e1cf3a4ddab3183dbeeabab9213"} Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.173500 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89cw6" event={"ID":"eacf2231-6eae-4b3c-aaea-ee0d90c4a150","Type":"ContainerDied","Data":"7c5cf4a9c0bf1968e0082ccafa3b7a58517384b9630fdc411db5f6acee453972"} Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.173546 4954 scope.go:117] "RemoveContainer" containerID="b1e17332acffcfd89f8db2a0c9f46aeb96232f2da01614b866f1ed4a97db9c21" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.173824 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-89cw6" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.182565 4954 generic.go:334] "Generic (PLEG): container finished" podID="4084f661-2955-43ba-a2d4-e6389830d100" containerID="35cdf07ce6733998fff22a56eb25f59a8ba9b435366fa01e6aac64db13999372" exitCode=0 Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.182705 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7x66v" event={"ID":"4084f661-2955-43ba-a2d4-e6389830d100","Type":"ContainerDied","Data":"35cdf07ce6733998fff22a56eb25f59a8ba9b435366fa01e6aac64db13999372"} Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.192657 4954 generic.go:334] "Generic (PLEG): container finished" podID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerID="72df4cc2d6b4b84c5cb6dfc9d53e848945487680f3f16efd084b2850dffea7f9" exitCode=0 Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.192730 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45kb2" event={"ID":"edc2fc95-fd10-469e-9e70-8368b75d1d82","Type":"ContainerDied","Data":"72df4cc2d6b4b84c5cb6dfc9d53e848945487680f3f16efd084b2850dffea7f9"} Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.194395 4954 generic.go:334] "Generic (PLEG): container finished" podID="6ff181e0-303d-4c55-9fc5-8c80561714ca" containerID="33195b1264b7a3fb8bf839bf88ae5b95c78bae9ae678023058d9a2aee9eab833" exitCode=0 Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.195128 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" event={"ID":"6ff181e0-303d-4c55-9fc5-8c80561714ca","Type":"ContainerDied","Data":"33195b1264b7a3fb8bf839bf88ae5b95c78bae9ae678023058d9a2aee9eab833"} Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.246562 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.289480 4954 scope.go:117] "RemoveContainer" containerID="07bd93f1f28866b091c0bd124fce07b4df961f2810726c9eece5724556ccd997" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.301303 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-89cw6"] Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.304689 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-89cw6"] Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.337336 4954 scope.go:117] "RemoveContainer" containerID="c2a1bff4736a10e68192e59f51324e1edcc4b33851f0ae5e7ccca4c4c92eb015" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.338023 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.449925 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-utilities\") pod \"da66ed1d-0022-4970-bc4c-aa6a62df0755\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.450045 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-catalog-content\") pod \"da66ed1d-0022-4970-bc4c-aa6a62df0755\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.450227 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv5dh\" (UniqueName: \"kubernetes.io/projected/da66ed1d-0022-4970-bc4c-aa6a62df0755-kube-api-access-hv5dh\") pod \"da66ed1d-0022-4970-bc4c-aa6a62df0755\" (UID: \"da66ed1d-0022-4970-bc4c-aa6a62df0755\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.450562 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.451702 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-utilities" (OuterVolumeSpecName: "utilities") pod "da66ed1d-0022-4970-bc4c-aa6a62df0755" (UID: "da66ed1d-0022-4970-bc4c-aa6a62df0755"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.453316 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.463458 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.464238 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da66ed1d-0022-4970-bc4c-aa6a62df0755-kube-api-access-hv5dh" (OuterVolumeSpecName: "kube-api-access-hv5dh") pod "da66ed1d-0022-4970-bc4c-aa6a62df0755" (UID: "da66ed1d-0022-4970-bc4c-aa6a62df0755"). InnerVolumeSpecName "kube-api-access-hv5dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.468336 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.472551 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.525117 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da66ed1d-0022-4970-bc4c-aa6a62df0755" (UID: "da66ed1d-0022-4970-bc4c-aa6a62df0755"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.556348 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-utilities\") pod \"edc2fc95-fd10-469e-9e70-8368b75d1d82\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.556658 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lhwm\" (UniqueName: \"kubernetes.io/projected/30f323d5-cbe3-4209-a708-a2e277eab190-kube-api-access-2lhwm\") pod \"30f323d5-cbe3-4209-a708-a2e277eab190\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.556796 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-catalog-content\") pod \"30f323d5-cbe3-4209-a708-a2e277eab190\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.556864 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-catalog-content\") pod \"edc2fc95-fd10-469e-9e70-8368b75d1d82\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.556945 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-trusted-ca\") pod \"6ff181e0-303d-4c55-9fc5-8c80561714ca\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.556975 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-utilities\") pod \"30f323d5-cbe3-4209-a708-a2e277eab190\" (UID: \"30f323d5-cbe3-4209-a708-a2e277eab190\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.557028 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-utilities\") pod \"4084f661-2955-43ba-a2d4-e6389830d100\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.557061 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhs5q\" (UniqueName: \"kubernetes.io/projected/6ff181e0-303d-4c55-9fc5-8c80561714ca-kube-api-access-dhs5q\") pod \"6ff181e0-303d-4c55-9fc5-8c80561714ca\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.557119 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhrdr\" (UniqueName: \"kubernetes.io/projected/4084f661-2955-43ba-a2d4-e6389830d100-kube-api-access-rhrdr\") pod \"4084f661-2955-43ba-a2d4-e6389830d100\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.557152 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct2v2\" (UniqueName: \"kubernetes.io/projected/edc2fc95-fd10-469e-9e70-8368b75d1d82-kube-api-access-ct2v2\") pod \"edc2fc95-fd10-469e-9e70-8368b75d1d82\" (UID: \"edc2fc95-fd10-469e-9e70-8368b75d1d82\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.557205 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-catalog-content\") pod \"4084f661-2955-43ba-a2d4-e6389830d100\" (UID: \"4084f661-2955-43ba-a2d4-e6389830d100\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.557234 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-operator-metrics\") pod \"6ff181e0-303d-4c55-9fc5-8c80561714ca\" (UID: \"6ff181e0-303d-4c55-9fc5-8c80561714ca\") " Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.557767 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da66ed1d-0022-4970-bc4c-aa6a62df0755-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.557789 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv5dh\" (UniqueName: \"kubernetes.io/projected/da66ed1d-0022-4970-bc4c-aa6a62df0755-kube-api-access-hv5dh\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.558066 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-utilities" (OuterVolumeSpecName: "utilities") pod "30f323d5-cbe3-4209-a708-a2e277eab190" (UID: "30f323d5-cbe3-4209-a708-a2e277eab190"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.560292 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-utilities" (OuterVolumeSpecName: "utilities") pod "edc2fc95-fd10-469e-9e70-8368b75d1d82" (UID: "edc2fc95-fd10-469e-9e70-8368b75d1d82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.560299 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30f323d5-cbe3-4209-a708-a2e277eab190-kube-api-access-2lhwm" (OuterVolumeSpecName: "kube-api-access-2lhwm") pod "30f323d5-cbe3-4209-a708-a2e277eab190" (UID: "30f323d5-cbe3-4209-a708-a2e277eab190"). InnerVolumeSpecName "kube-api-access-2lhwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.561613 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "6ff181e0-303d-4c55-9fc5-8c80561714ca" (UID: "6ff181e0-303d-4c55-9fc5-8c80561714ca"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.561720 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "6ff181e0-303d-4c55-9fc5-8c80561714ca" (UID: "6ff181e0-303d-4c55-9fc5-8c80561714ca"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.562794 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4084f661-2955-43ba-a2d4-e6389830d100-kube-api-access-rhrdr" (OuterVolumeSpecName: "kube-api-access-rhrdr") pod "4084f661-2955-43ba-a2d4-e6389830d100" (UID: "4084f661-2955-43ba-a2d4-e6389830d100"). InnerVolumeSpecName "kube-api-access-rhrdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.562926 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-utilities" (OuterVolumeSpecName: "utilities") pod "4084f661-2955-43ba-a2d4-e6389830d100" (UID: "4084f661-2955-43ba-a2d4-e6389830d100"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.566587 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edc2fc95-fd10-469e-9e70-8368b75d1d82-kube-api-access-ct2v2" (OuterVolumeSpecName: "kube-api-access-ct2v2") pod "edc2fc95-fd10-469e-9e70-8368b75d1d82" (UID: "edc2fc95-fd10-469e-9e70-8368b75d1d82"). InnerVolumeSpecName "kube-api-access-ct2v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.566812 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ff181e0-303d-4c55-9fc5-8c80561714ca-kube-api-access-dhs5q" (OuterVolumeSpecName: "kube-api-access-dhs5q") pod "6ff181e0-303d-4c55-9fc5-8c80561714ca" (UID: "6ff181e0-303d-4c55-9fc5-8c80561714ca"). InnerVolumeSpecName "kube-api-access-dhs5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.592386 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4084f661-2955-43ba-a2d4-e6389830d100" (UID: "4084f661-2955-43ba-a2d4-e6389830d100"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.631890 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "edc2fc95-fd10-469e-9e70-8368b75d1d82" (UID: "edc2fc95-fd10-469e-9e70-8368b75d1d82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.659981 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.660026 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhs5q\" (UniqueName: \"kubernetes.io/projected/6ff181e0-303d-4c55-9fc5-8c80561714ca-kube-api-access-dhs5q\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.660044 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhrdr\" (UniqueName: \"kubernetes.io/projected/4084f661-2955-43ba-a2d4-e6389830d100-kube-api-access-rhrdr\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.660057 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct2v2\" (UniqueName: \"kubernetes.io/projected/edc2fc95-fd10-469e-9e70-8368b75d1d82-kube-api-access-ct2v2\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.660069 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4084f661-2955-43ba-a2d4-e6389830d100-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.660082 4954 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.660095 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.660107 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lhwm\" (UniqueName: \"kubernetes.io/projected/30f323d5-cbe3-4209-a708-a2e277eab190-kube-api-access-2lhwm\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.660118 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edc2fc95-fd10-469e-9e70-8368b75d1d82-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.660130 4954 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ff181e0-303d-4c55-9fc5-8c80561714ca-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.660141 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.694723 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30f323d5-cbe3-4209-a708-a2e277eab190" (UID: "30f323d5-cbe3-4209-a708-a2e277eab190"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.714203 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-td9wp"] Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.764061 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30f323d5-cbe3-4209-a708-a2e277eab190-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:14 crc kubenswrapper[4954]: I1209 17:12:14.848639 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4svl9"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.204332 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4cpr" event={"ID":"da66ed1d-0022-4970-bc4c-aa6a62df0755","Type":"ContainerDied","Data":"60289ebc00941a667f985407b5d50918c8910be7d5a67f895e353a9ea36001d5"} Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.204373 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4cpr" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.204404 4954 scope.go:117] "RemoveContainer" containerID="697f4a455885134b6769d2283d6a950e72e67e1cf3a4ddab3183dbeeabab9213" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.212234 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7x66v" event={"ID":"4084f661-2955-43ba-a2d4-e6389830d100","Type":"ContainerDied","Data":"f91cfe648e0d4d35425d69aaebc9be4a68610ed3503f81a6556322492d38f750"} Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.212920 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7x66v" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.216617 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-45kb2" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.216656 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-45kb2" event={"ID":"edc2fc95-fd10-469e-9e70-8368b75d1d82","Type":"ContainerDied","Data":"8ccd1e4a1052f69b487e4679d4d6154d700e592b261ff1d058c0521a9c7fcf76"} Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.221238 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" event={"ID":"6ff181e0-303d-4c55-9fc5-8c80561714ca","Type":"ContainerDied","Data":"e91f29c426f2f85f03f8352dfdf3843dde12af4e3d1d0cdc91538a69c3c97044"} Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.221288 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v98r2" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.223791 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" event={"ID":"2fd60f62-8fed-4f13-9da7-926b99b7d4aa","Type":"ContainerStarted","Data":"9d258e7a77a41f9be398f9327881a7698d3780823ce3283931e4a78c6c5cf4e7"} Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.228883 4954 scope.go:117] "RemoveContainer" containerID="01251ec2969f4ee6b355ba129fd60509016f6c24fe5f9f7cbf928ce6575d30c3" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.229680 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8lch" event={"ID":"30f323d5-cbe3-4209-a708-a2e277eab190","Type":"ContainerDied","Data":"6b7b8c08980f4a18b9830b023a1a4630d69f0ac5b2becf62a2d4b5f6ba375fb8"} Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.229733 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8lch" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.230076 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-td9wp" podUID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerName="registry-server" containerID="cri-o://dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee" gracePeriod=30 Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.258841 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g4cpr"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.271923 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g4cpr"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.277299 4954 scope.go:117] "RemoveContainer" containerID="2ba8d8af2a88d2314467f4d684e5e8381510d4990cdf2e34256ce7e0be366c1f" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.281279 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-45kb2"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.287550 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-45kb2"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.311086 4954 scope.go:117] "RemoveContainer" containerID="35cdf07ce6733998fff22a56eb25f59a8ba9b435366fa01e6aac64db13999372" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.311086 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v98r2"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.321415 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v98r2"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.327628 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7x66v"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.343691 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7x66v"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.344174 4954 scope.go:117] "RemoveContainer" containerID="d106d7667edbb1b61e291ea928a889a5ac8bb54f728fc55c64fd660a671061ba" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.347927 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8lch"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.351880 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q8lch"] Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.362181 4954 scope.go:117] "RemoveContainer" containerID="109e5810b1a657ab4402f83b2e1eaf0cc9eb8ba389364f03b5eba3c680f0b95c" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.378324 4954 scope.go:117] "RemoveContainer" containerID="72df4cc2d6b4b84c5cb6dfc9d53e848945487680f3f16efd084b2850dffea7f9" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.395003 4954 scope.go:117] "RemoveContainer" containerID="ee355a22239b8c9839879a4b35a6354ba052d7c9447f43f3356fc47676aa3744" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.413198 4954 scope.go:117] "RemoveContainer" containerID="ff2f8a68fa23172a3d3184ea750e5a76bc4e281859d536e859426013b7503cc0" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.429909 4954 scope.go:117] "RemoveContainer" containerID="33195b1264b7a3fb8bf839bf88ae5b95c78bae9ae678023058d9a2aee9eab833" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.450018 4954 scope.go:117] "RemoveContainer" containerID="9289442f940546cbcd543822d629ae49e411d9d8eb95f4dde17e9bebf5ff4156" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.466653 4954 scope.go:117] "RemoveContainer" containerID="be44a0e2910f1dd296bf694436bb1194a03295ced9176724b66b2d24fabacb6a" Dec 09 17:12:15 crc kubenswrapper[4954]: I1209 17:12:15.548380 4954 scope.go:117] "RemoveContainer" containerID="83436d1245c55f2a5637d35967cf5b613ca849dce49eb02613142b119b520899" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.131429 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30f323d5-cbe3-4209-a708-a2e277eab190" path="/var/lib/kubelet/pods/30f323d5-cbe3-4209-a708-a2e277eab190/volumes" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.132152 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4084f661-2955-43ba-a2d4-e6389830d100" path="/var/lib/kubelet/pods/4084f661-2955-43ba-a2d4-e6389830d100/volumes" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.132887 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ff181e0-303d-4c55-9fc5-8c80561714ca" path="/var/lib/kubelet/pods/6ff181e0-303d-4c55-9fc5-8c80561714ca/volumes" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.134845 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da66ed1d-0022-4970-bc4c-aa6a62df0755" path="/var/lib/kubelet/pods/da66ed1d-0022-4970-bc4c-aa6a62df0755/volumes" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.135516 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eacf2231-6eae-4b3c-aaea-ee0d90c4a150" path="/var/lib/kubelet/pods/eacf2231-6eae-4b3c-aaea-ee0d90c4a150/volumes" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.137093 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edc2fc95-fd10-469e-9e70-8368b75d1d82" path="/var/lib/kubelet/pods/edc2fc95-fd10-469e-9e70-8368b75d1d82/volumes" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138087 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2jsfq"] Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138403 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerName="extract-content" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138431 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerName="extract-content" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138447 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30f323d5-cbe3-4209-a708-a2e277eab190" containerName="extract-utilities" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138456 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="30f323d5-cbe3-4209-a708-a2e277eab190" containerName="extract-utilities" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138465 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4084f661-2955-43ba-a2d4-e6389830d100" containerName="extract-content" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138472 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="4084f661-2955-43ba-a2d4-e6389830d100" containerName="extract-content" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138485 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerName="extract-utilities" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138493 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerName="extract-utilities" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138502 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138510 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138523 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4084f661-2955-43ba-a2d4-e6389830d100" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138532 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="4084f661-2955-43ba-a2d4-e6389830d100" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138546 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerName="extract-content" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138556 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerName="extract-content" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138566 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4084f661-2955-43ba-a2d4-e6389830d100" containerName="extract-utilities" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138575 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="4084f661-2955-43ba-a2d4-e6389830d100" containerName="extract-utilities" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138590 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerName="extract-utilities" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138615 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerName="extract-utilities" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138632 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff181e0-303d-4c55-9fc5-8c80561714ca" containerName="marketplace-operator" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138640 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff181e0-303d-4c55-9fc5-8c80561714ca" containerName="marketplace-operator" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138650 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30f323d5-cbe3-4209-a708-a2e277eab190" containerName="extract-content" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138657 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="30f323d5-cbe3-4209-a708-a2e277eab190" containerName="extract-content" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138669 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30f323d5-cbe3-4209-a708-a2e277eab190" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138680 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="30f323d5-cbe3-4209-a708-a2e277eab190" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.138700 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138709 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138842 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="30f323d5-cbe3-4209-a708-a2e277eab190" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138853 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="da66ed1d-0022-4970-bc4c-aa6a62df0755" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138863 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="4084f661-2955-43ba-a2d4-e6389830d100" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138871 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="edc2fc95-fd10-469e-9e70-8368b75d1d82" containerName="registry-server" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.138881 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff181e0-303d-4c55-9fc5-8c80561714ca" containerName="marketplace-operator" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.140384 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.144016 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.149149 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2jsfq"] Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.160664 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.197328 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-utilities\") pod \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.197438 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-catalog-content\") pod \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.197615 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgctk\" (UniqueName: \"kubernetes.io/projected/e37a9f7e-c431-4ee3-8286-4716a36f1f63-kube-api-access-mgctk\") pod \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\" (UID: \"e37a9f7e-c431-4ee3-8286-4716a36f1f63\") " Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.198052 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-catalog-content\") pod \"community-operators-2jsfq\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.198101 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-utilities\") pod \"community-operators-2jsfq\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.198155 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2ssq\" (UniqueName: \"kubernetes.io/projected/9ef7858f-b8fb-429b-886b-3671b1ab191e-kube-api-access-x2ssq\") pod \"community-operators-2jsfq\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.198873 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-utilities" (OuterVolumeSpecName: "utilities") pod "e37a9f7e-c431-4ee3-8286-4716a36f1f63" (UID: "e37a9f7e-c431-4ee3-8286-4716a36f1f63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.211042 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e37a9f7e-c431-4ee3-8286-4716a36f1f63-kube-api-access-mgctk" (OuterVolumeSpecName: "kube-api-access-mgctk") pod "e37a9f7e-c431-4ee3-8286-4716a36f1f63" (UID: "e37a9f7e-c431-4ee3-8286-4716a36f1f63"). InnerVolumeSpecName "kube-api-access-mgctk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.249140 4954 generic.go:334] "Generic (PLEG): container finished" podID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerID="dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee" exitCode=0 Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.249304 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-td9wp" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.249331 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9wp" event={"ID":"e37a9f7e-c431-4ee3-8286-4716a36f1f63","Type":"ContainerDied","Data":"dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee"} Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.250075 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9wp" event={"ID":"e37a9f7e-c431-4ee3-8286-4716a36f1f63","Type":"ContainerDied","Data":"8360cb043f056f63f2d7c8446b40b5618a1458f14b8f389e3f0343b105e6096a"} Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.250136 4954 scope.go:117] "RemoveContainer" containerID="dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.259781 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" event={"ID":"2fd60f62-8fed-4f13-9da7-926b99b7d4aa","Type":"ContainerStarted","Data":"53607d1a3d89ad8a3906af2fbe384578a91ce768bcebd077d46bd1965ccd1b88"} Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.259870 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.261087 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e37a9f7e-c431-4ee3-8286-4716a36f1f63" (UID: "e37a9f7e-c431-4ee3-8286-4716a36f1f63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.264306 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.286210 4954 scope.go:117] "RemoveContainer" containerID="c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.288507 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4svl9" podStartSLOduration=3.288484794 podStartE2EDuration="3.288484794s" podCreationTimestamp="2025-12-09 17:12:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:12:16.277979826 +0000 UTC m=+932.666153656" watchObservedRunningTime="2025-12-09 17:12:16.288484794 +0000 UTC m=+932.676658614" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.299046 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-utilities\") pod \"community-operators-2jsfq\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.299099 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2ssq\" (UniqueName: \"kubernetes.io/projected/9ef7858f-b8fb-429b-886b-3671b1ab191e-kube-api-access-x2ssq\") pod \"community-operators-2jsfq\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.299219 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-catalog-content\") pod \"community-operators-2jsfq\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.299263 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgctk\" (UniqueName: \"kubernetes.io/projected/e37a9f7e-c431-4ee3-8286-4716a36f1f63-kube-api-access-mgctk\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.299278 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.299290 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37a9f7e-c431-4ee3-8286-4716a36f1f63-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.299503 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-utilities\") pod \"community-operators-2jsfq\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.299557 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-catalog-content\") pod \"community-operators-2jsfq\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.315758 4954 scope.go:117] "RemoveContainer" containerID="d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.325834 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2ssq\" (UniqueName: \"kubernetes.io/projected/9ef7858f-b8fb-429b-886b-3671b1ab191e-kube-api-access-x2ssq\") pod \"community-operators-2jsfq\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.349550 4954 scope.go:117] "RemoveContainer" containerID="dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.350863 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee\": container with ID starting with dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee not found: ID does not exist" containerID="dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.350898 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee"} err="failed to get container status \"dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee\": rpc error: code = NotFound desc = could not find container \"dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee\": container with ID starting with dadc4335eb6a294094196575143d95886861d35105bdd6ff6ed08d43e8f6f7ee not found: ID does not exist" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.350920 4954 scope.go:117] "RemoveContainer" containerID="c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.351314 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3\": container with ID starting with c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3 not found: ID does not exist" containerID="c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.351337 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3"} err="failed to get container status \"c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3\": rpc error: code = NotFound desc = could not find container \"c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3\": container with ID starting with c3b883f820cbdb724129a595c83e3798a569cd417f6678162e7fff91577e21f3 not found: ID does not exist" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.351355 4954 scope.go:117] "RemoveContainer" containerID="d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18" Dec 09 17:12:16 crc kubenswrapper[4954]: E1209 17:12:16.351721 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18\": container with ID starting with d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18 not found: ID does not exist" containerID="d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.351778 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18"} err="failed to get container status \"d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18\": rpc error: code = NotFound desc = could not find container \"d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18\": container with ID starting with d5a9819a6e25ddabaea0db0ff184c7b96abd125a4e10f74c4c1ae106f3844a18 not found: ID does not exist" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.482732 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.600802 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-td9wp"] Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.612670 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-td9wp"] Dec 09 17:12:16 crc kubenswrapper[4954]: I1209 17:12:16.723309 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2jsfq"] Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.124473 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-krpqr"] Dec 09 17:12:17 crc kubenswrapper[4954]: E1209 17:12:17.124869 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerName="extract-utilities" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.124891 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerName="extract-utilities" Dec 09 17:12:17 crc kubenswrapper[4954]: E1209 17:12:17.124907 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerName="extract-content" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.124914 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerName="extract-content" Dec 09 17:12:17 crc kubenswrapper[4954]: E1209 17:12:17.124927 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerName="registry-server" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.124935 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerName="registry-server" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.125082 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" containerName="registry-server" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.126291 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.132518 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.142715 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-krpqr"] Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.215714 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbbrn\" (UniqueName: \"kubernetes.io/projected/315ebcec-7a31-4e51-9f9b-46f715643fe5-kube-api-access-nbbrn\") pod \"redhat-marketplace-krpqr\" (UID: \"315ebcec-7a31-4e51-9f9b-46f715643fe5\") " pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.216189 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315ebcec-7a31-4e51-9f9b-46f715643fe5-catalog-content\") pod \"redhat-marketplace-krpqr\" (UID: \"315ebcec-7a31-4e51-9f9b-46f715643fe5\") " pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.216273 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315ebcec-7a31-4e51-9f9b-46f715643fe5-utilities\") pod \"redhat-marketplace-krpqr\" (UID: \"315ebcec-7a31-4e51-9f9b-46f715643fe5\") " pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.284339 4954 generic.go:334] "Generic (PLEG): container finished" podID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerID="3993e73bf7b5be3995b3cdc2f27835f3d9dd3f579f1ac385fa589059ad803d72" exitCode=0 Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.284419 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jsfq" event={"ID":"9ef7858f-b8fb-429b-886b-3671b1ab191e","Type":"ContainerDied","Data":"3993e73bf7b5be3995b3cdc2f27835f3d9dd3f579f1ac385fa589059ad803d72"} Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.284871 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jsfq" event={"ID":"9ef7858f-b8fb-429b-886b-3671b1ab191e","Type":"ContainerStarted","Data":"5f8bd0e13d24ac263c41cef14987488d6f49d4a978a37e78c0452b4303cee0b5"} Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.319573 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315ebcec-7a31-4e51-9f9b-46f715643fe5-catalog-content\") pod \"redhat-marketplace-krpqr\" (UID: \"315ebcec-7a31-4e51-9f9b-46f715643fe5\") " pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.319655 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315ebcec-7a31-4e51-9f9b-46f715643fe5-utilities\") pod \"redhat-marketplace-krpqr\" (UID: \"315ebcec-7a31-4e51-9f9b-46f715643fe5\") " pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.319890 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315ebcec-7a31-4e51-9f9b-46f715643fe5-catalog-content\") pod \"redhat-marketplace-krpqr\" (UID: \"315ebcec-7a31-4e51-9f9b-46f715643fe5\") " pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.320734 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbbrn\" (UniqueName: \"kubernetes.io/projected/315ebcec-7a31-4e51-9f9b-46f715643fe5-kube-api-access-nbbrn\") pod \"redhat-marketplace-krpqr\" (UID: \"315ebcec-7a31-4e51-9f9b-46f715643fe5\") " pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.320001 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315ebcec-7a31-4e51-9f9b-46f715643fe5-utilities\") pod \"redhat-marketplace-krpqr\" (UID: \"315ebcec-7a31-4e51-9f9b-46f715643fe5\") " pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.346990 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbbrn\" (UniqueName: \"kubernetes.io/projected/315ebcec-7a31-4e51-9f9b-46f715643fe5-kube-api-access-nbbrn\") pod \"redhat-marketplace-krpqr\" (UID: \"315ebcec-7a31-4e51-9f9b-46f715643fe5\") " pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.445524 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:17 crc kubenswrapper[4954]: I1209 17:12:17.673142 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-krpqr"] Dec 09 17:12:17 crc kubenswrapper[4954]: W1209 17:12:17.680775 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod315ebcec_7a31_4e51_9f9b_46f715643fe5.slice/crio-d901dd4a7227e77b457281b9a6a66752de225b67389b848966bbefeb30bb2509 WatchSource:0}: Error finding container d901dd4a7227e77b457281b9a6a66752de225b67389b848966bbefeb30bb2509: Status 404 returned error can't find the container with id d901dd4a7227e77b457281b9a6a66752de225b67389b848966bbefeb30bb2509 Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.131217 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e37a9f7e-c431-4ee3-8286-4716a36f1f63" path="/var/lib/kubelet/pods/e37a9f7e-c431-4ee3-8286-4716a36f1f63/volumes" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.299544 4954 generic.go:334] "Generic (PLEG): container finished" podID="315ebcec-7a31-4e51-9f9b-46f715643fe5" containerID="1796955a9faf42b5d4bad9ba17b6cdb34b80f8c9b3823c52751f654a307fa66a" exitCode=0 Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.300771 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krpqr" event={"ID":"315ebcec-7a31-4e51-9f9b-46f715643fe5","Type":"ContainerDied","Data":"1796955a9faf42b5d4bad9ba17b6cdb34b80f8c9b3823c52751f654a307fa66a"} Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.300826 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krpqr" event={"ID":"315ebcec-7a31-4e51-9f9b-46f715643fe5","Type":"ContainerStarted","Data":"d901dd4a7227e77b457281b9a6a66752de225b67389b848966bbefeb30bb2509"} Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.524805 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gg7lf"] Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.526682 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.531426 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.547788 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gg7lf"] Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.641071 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da760521-0631-4573-a06c-dc5bbf6dddc7-utilities\") pod \"redhat-operators-gg7lf\" (UID: \"da760521-0631-4573-a06c-dc5bbf6dddc7\") " pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.641166 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da760521-0631-4573-a06c-dc5bbf6dddc7-catalog-content\") pod \"redhat-operators-gg7lf\" (UID: \"da760521-0631-4573-a06c-dc5bbf6dddc7\") " pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.641192 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5c42\" (UniqueName: \"kubernetes.io/projected/da760521-0631-4573-a06c-dc5bbf6dddc7-kube-api-access-w5c42\") pod \"redhat-operators-gg7lf\" (UID: \"da760521-0631-4573-a06c-dc5bbf6dddc7\") " pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.742149 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da760521-0631-4573-a06c-dc5bbf6dddc7-utilities\") pod \"redhat-operators-gg7lf\" (UID: \"da760521-0631-4573-a06c-dc5bbf6dddc7\") " pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.742290 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da760521-0631-4573-a06c-dc5bbf6dddc7-catalog-content\") pod \"redhat-operators-gg7lf\" (UID: \"da760521-0631-4573-a06c-dc5bbf6dddc7\") " pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.742326 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5c42\" (UniqueName: \"kubernetes.io/projected/da760521-0631-4573-a06c-dc5bbf6dddc7-kube-api-access-w5c42\") pod \"redhat-operators-gg7lf\" (UID: \"da760521-0631-4573-a06c-dc5bbf6dddc7\") " pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.742808 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da760521-0631-4573-a06c-dc5bbf6dddc7-utilities\") pod \"redhat-operators-gg7lf\" (UID: \"da760521-0631-4573-a06c-dc5bbf6dddc7\") " pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.743024 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da760521-0631-4573-a06c-dc5bbf6dddc7-catalog-content\") pod \"redhat-operators-gg7lf\" (UID: \"da760521-0631-4573-a06c-dc5bbf6dddc7\") " pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.768186 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5c42\" (UniqueName: \"kubernetes.io/projected/da760521-0631-4573-a06c-dc5bbf6dddc7-kube-api-access-w5c42\") pod \"redhat-operators-gg7lf\" (UID: \"da760521-0631-4573-a06c-dc5bbf6dddc7\") " pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:18 crc kubenswrapper[4954]: I1209 17:12:18.843918 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.311491 4954 generic.go:334] "Generic (PLEG): container finished" podID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerID="157d27832ca42157552f6cb23c06586a2ea392db8ee12278155142ed7712aab2" exitCode=0 Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.311625 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jsfq" event={"ID":"9ef7858f-b8fb-429b-886b-3671b1ab191e","Type":"ContainerDied","Data":"157d27832ca42157552f6cb23c06586a2ea392db8ee12278155142ed7712aab2"} Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.320978 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gg7lf"] Dec 09 17:12:19 crc kubenswrapper[4954]: W1209 17:12:19.330902 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda760521_0631_4573_a06c_dc5bbf6dddc7.slice/crio-16ab7efb209112ce577591f6ba905ec0d9f96ae1471783614ecf7f5fffef0b92 WatchSource:0}: Error finding container 16ab7efb209112ce577591f6ba905ec0d9f96ae1471783614ecf7f5fffef0b92: Status 404 returned error can't find the container with id 16ab7efb209112ce577591f6ba905ec0d9f96ae1471783614ecf7f5fffef0b92 Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.524132 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t9mg8"] Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.526575 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.532989 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.539098 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t9mg8"] Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.559137 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbvnn\" (UniqueName: \"kubernetes.io/projected/98dfad17-db42-4215-add1-3d3b01a971b0-kube-api-access-bbvnn\") pod \"certified-operators-t9mg8\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.559217 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-catalog-content\") pod \"certified-operators-t9mg8\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.559261 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-utilities\") pod \"certified-operators-t9mg8\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.661055 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbvnn\" (UniqueName: \"kubernetes.io/projected/98dfad17-db42-4215-add1-3d3b01a971b0-kube-api-access-bbvnn\") pod \"certified-operators-t9mg8\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.661139 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-catalog-content\") pod \"certified-operators-t9mg8\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.661177 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-utilities\") pod \"certified-operators-t9mg8\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.661812 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-utilities\") pod \"certified-operators-t9mg8\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.661936 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-catalog-content\") pod \"certified-operators-t9mg8\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.690097 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbvnn\" (UniqueName: \"kubernetes.io/projected/98dfad17-db42-4215-add1-3d3b01a971b0-kube-api-access-bbvnn\") pod \"certified-operators-t9mg8\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:19 crc kubenswrapper[4954]: I1209 17:12:19.849686 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:20 crc kubenswrapper[4954]: I1209 17:12:20.279802 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t9mg8"] Dec 09 17:12:20 crc kubenswrapper[4954]: W1209 17:12:20.297932 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98dfad17_db42_4215_add1_3d3b01a971b0.slice/crio-9443228bff0f5a04c931e3b14c888d10f09913c7edafc3856936ebf25e853bc6 WatchSource:0}: Error finding container 9443228bff0f5a04c931e3b14c888d10f09913c7edafc3856936ebf25e853bc6: Status 404 returned error can't find the container with id 9443228bff0f5a04c931e3b14c888d10f09913c7edafc3856936ebf25e853bc6 Dec 09 17:12:20 crc kubenswrapper[4954]: I1209 17:12:20.319878 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9mg8" event={"ID":"98dfad17-db42-4215-add1-3d3b01a971b0","Type":"ContainerStarted","Data":"9443228bff0f5a04c931e3b14c888d10f09913c7edafc3856936ebf25e853bc6"} Dec 09 17:12:20 crc kubenswrapper[4954]: I1209 17:12:20.320903 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg7lf" event={"ID":"da760521-0631-4573-a06c-dc5bbf6dddc7","Type":"ContainerStarted","Data":"16ab7efb209112ce577591f6ba905ec0d9f96ae1471783614ecf7f5fffef0b92"} Dec 09 17:12:22 crc kubenswrapper[4954]: I1209 17:12:22.338291 4954 generic.go:334] "Generic (PLEG): container finished" podID="da760521-0631-4573-a06c-dc5bbf6dddc7" containerID="42eb4f59034de466e5bbbd77fbbb682d1622eca408b3be7d6475b7fecb5a0ef7" exitCode=0 Dec 09 17:12:22 crc kubenswrapper[4954]: I1209 17:12:22.338407 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg7lf" event={"ID":"da760521-0631-4573-a06c-dc5bbf6dddc7","Type":"ContainerDied","Data":"42eb4f59034de466e5bbbd77fbbb682d1622eca408b3be7d6475b7fecb5a0ef7"} Dec 09 17:12:22 crc kubenswrapper[4954]: I1209 17:12:22.340825 4954 generic.go:334] "Generic (PLEG): container finished" podID="98dfad17-db42-4215-add1-3d3b01a971b0" containerID="014be430bee464ff6819affc7cf16f33194c0a2572a3c4cf9ed0f0c1f69f2354" exitCode=0 Dec 09 17:12:22 crc kubenswrapper[4954]: I1209 17:12:22.340940 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9mg8" event={"ID":"98dfad17-db42-4215-add1-3d3b01a971b0","Type":"ContainerDied","Data":"014be430bee464ff6819affc7cf16f33194c0a2572a3c4cf9ed0f0c1f69f2354"} Dec 09 17:12:22 crc kubenswrapper[4954]: I1209 17:12:22.344083 4954 generic.go:334] "Generic (PLEG): container finished" podID="315ebcec-7a31-4e51-9f9b-46f715643fe5" containerID="e0378ee8767e6d34446967ef2c82391568e1c95e399dfa21b66fed5cfb99bab8" exitCode=0 Dec 09 17:12:22 crc kubenswrapper[4954]: I1209 17:12:22.344143 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krpqr" event={"ID":"315ebcec-7a31-4e51-9f9b-46f715643fe5","Type":"ContainerDied","Data":"e0378ee8767e6d34446967ef2c82391568e1c95e399dfa21b66fed5cfb99bab8"} Dec 09 17:12:22 crc kubenswrapper[4954]: I1209 17:12:22.347289 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jsfq" event={"ID":"9ef7858f-b8fb-429b-886b-3671b1ab191e","Type":"ContainerStarted","Data":"e0d97a0477dd29addf0385ae26a2abeecbc332f4b210068fc1b6deb06096ea22"} Dec 09 17:12:22 crc kubenswrapper[4954]: I1209 17:12:22.397426 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2jsfq" podStartSLOduration=2.131596635 podStartE2EDuration="6.397404366s" podCreationTimestamp="2025-12-09 17:12:16 +0000 UTC" firstStartedPulling="2025-12-09 17:12:17.287187072 +0000 UTC m=+933.675360892" lastFinishedPulling="2025-12-09 17:12:21.552994803 +0000 UTC m=+937.941168623" observedRunningTime="2025-12-09 17:12:22.396019292 +0000 UTC m=+938.784193112" watchObservedRunningTime="2025-12-09 17:12:22.397404366 +0000 UTC m=+938.785578186" Dec 09 17:12:23 crc kubenswrapper[4954]: I1209 17:12:23.356328 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg7lf" event={"ID":"da760521-0631-4573-a06c-dc5bbf6dddc7","Type":"ContainerStarted","Data":"cf631e3fdaf1fed9d17884ae0b32943d5bff170980f5f91cefea2a29650e5b89"} Dec 09 17:12:23 crc kubenswrapper[4954]: I1209 17:12:23.358788 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9mg8" event={"ID":"98dfad17-db42-4215-add1-3d3b01a971b0","Type":"ContainerStarted","Data":"a01d965c56720dbf5f22bd97da4332e2aca6463852e930ea0d400c7034578267"} Dec 09 17:12:23 crc kubenswrapper[4954]: I1209 17:12:23.362705 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krpqr" event={"ID":"315ebcec-7a31-4e51-9f9b-46f715643fe5","Type":"ContainerStarted","Data":"9690491c9d09d0a256a80d654db5b87cbfd0c4b98f38d7fb90a023e911c2e1ce"} Dec 09 17:12:23 crc kubenswrapper[4954]: I1209 17:12:23.430316 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-krpqr" podStartSLOduration=1.750967788 podStartE2EDuration="6.430285953s" podCreationTimestamp="2025-12-09 17:12:17 +0000 UTC" firstStartedPulling="2025-12-09 17:12:18.304220394 +0000 UTC m=+934.692394214" lastFinishedPulling="2025-12-09 17:12:22.983538559 +0000 UTC m=+939.371712379" observedRunningTime="2025-12-09 17:12:23.423264863 +0000 UTC m=+939.811438703" watchObservedRunningTime="2025-12-09 17:12:23.430285953 +0000 UTC m=+939.818459773" Dec 09 17:12:24 crc kubenswrapper[4954]: I1209 17:12:24.369056 4954 generic.go:334] "Generic (PLEG): container finished" podID="da760521-0631-4573-a06c-dc5bbf6dddc7" containerID="cf631e3fdaf1fed9d17884ae0b32943d5bff170980f5f91cefea2a29650e5b89" exitCode=0 Dec 09 17:12:24 crc kubenswrapper[4954]: I1209 17:12:24.369123 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg7lf" event={"ID":"da760521-0631-4573-a06c-dc5bbf6dddc7","Type":"ContainerDied","Data":"cf631e3fdaf1fed9d17884ae0b32943d5bff170980f5f91cefea2a29650e5b89"} Dec 09 17:12:24 crc kubenswrapper[4954]: I1209 17:12:24.371650 4954 generic.go:334] "Generic (PLEG): container finished" podID="98dfad17-db42-4215-add1-3d3b01a971b0" containerID="a01d965c56720dbf5f22bd97da4332e2aca6463852e930ea0d400c7034578267" exitCode=0 Dec 09 17:12:24 crc kubenswrapper[4954]: I1209 17:12:24.371778 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9mg8" event={"ID":"98dfad17-db42-4215-add1-3d3b01a971b0","Type":"ContainerDied","Data":"a01d965c56720dbf5f22bd97da4332e2aca6463852e930ea0d400c7034578267"} Dec 09 17:12:26 crc kubenswrapper[4954]: I1209 17:12:26.389744 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg7lf" event={"ID":"da760521-0631-4573-a06c-dc5bbf6dddc7","Type":"ContainerStarted","Data":"c0a31672cd516decdc28886879122f779c1584c813cf39d50371cec0f7bb03d9"} Dec 09 17:12:26 crc kubenswrapper[4954]: I1209 17:12:26.392105 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9mg8" event={"ID":"98dfad17-db42-4215-add1-3d3b01a971b0","Type":"ContainerStarted","Data":"01ebf81616168a4141040377a1ad3c451c3b64012972858a42beed2824510efe"} Dec 09 17:12:26 crc kubenswrapper[4954]: I1209 17:12:26.415515 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gg7lf" podStartSLOduration=5.34224566 podStartE2EDuration="8.415495088s" podCreationTimestamp="2025-12-09 17:12:18 +0000 UTC" firstStartedPulling="2025-12-09 17:12:22.340613379 +0000 UTC m=+938.728787199" lastFinishedPulling="2025-12-09 17:12:25.413862807 +0000 UTC m=+941.802036627" observedRunningTime="2025-12-09 17:12:26.410708687 +0000 UTC m=+942.798882517" watchObservedRunningTime="2025-12-09 17:12:26.415495088 +0000 UTC m=+942.803668908" Dec 09 17:12:26 crc kubenswrapper[4954]: I1209 17:12:26.430920 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t9mg8" podStartSLOduration=4.387016489 podStartE2EDuration="7.430865298s" podCreationTimestamp="2025-12-09 17:12:19 +0000 UTC" firstStartedPulling="2025-12-09 17:12:22.343269382 +0000 UTC m=+938.731443202" lastFinishedPulling="2025-12-09 17:12:25.387118191 +0000 UTC m=+941.775292011" observedRunningTime="2025-12-09 17:12:26.428617418 +0000 UTC m=+942.816791258" watchObservedRunningTime="2025-12-09 17:12:26.430865298 +0000 UTC m=+942.819039118" Dec 09 17:12:26 crc kubenswrapper[4954]: I1209 17:12:26.483403 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:26 crc kubenswrapper[4954]: I1209 17:12:26.484136 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:26 crc kubenswrapper[4954]: I1209 17:12:26.536789 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:27 crc kubenswrapper[4954]: I1209 17:12:27.440457 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2jsfq" Dec 09 17:12:27 crc kubenswrapper[4954]: I1209 17:12:27.446075 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:27 crc kubenswrapper[4954]: I1209 17:12:27.446173 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:27 crc kubenswrapper[4954]: I1209 17:12:27.493915 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:28 crc kubenswrapper[4954]: I1209 17:12:28.457047 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-krpqr" Dec 09 17:12:28 crc kubenswrapper[4954]: I1209 17:12:28.844992 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:28 crc kubenswrapper[4954]: I1209 17:12:28.845442 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:29 crc kubenswrapper[4954]: I1209 17:12:29.850031 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:29 crc kubenswrapper[4954]: I1209 17:12:29.850087 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:29 crc kubenswrapper[4954]: I1209 17:12:29.890094 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gg7lf" podUID="da760521-0631-4573-a06c-dc5bbf6dddc7" containerName="registry-server" probeResult="failure" output=< Dec 09 17:12:29 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 17:12:29 crc kubenswrapper[4954]: > Dec 09 17:12:29 crc kubenswrapper[4954]: I1209 17:12:29.902782 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:30 crc kubenswrapper[4954]: I1209 17:12:30.472697 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:12:38 crc kubenswrapper[4954]: I1209 17:12:38.883303 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:38 crc kubenswrapper[4954]: I1209 17:12:38.929632 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gg7lf" Dec 09 17:12:44 crc kubenswrapper[4954]: I1209 17:12:44.274030 4954 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","podeacf2231-6eae-4b3c-aaea-ee0d90c4a150"] err="unable to destroy cgroup paths for cgroup [kubepods burstable podeacf2231-6eae-4b3c-aaea-ee0d90c4a150] : Timed out while waiting for systemd to remove kubepods-burstable-podeacf2231_6eae_4b3c_aaea_ee0d90c4a150.slice" Dec 09 17:13:13 crc kubenswrapper[4954]: I1209 17:13:13.755830 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:13:13 crc kubenswrapper[4954]: I1209 17:13:13.756544 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:13:43 crc kubenswrapper[4954]: I1209 17:13:43.755502 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:13:43 crc kubenswrapper[4954]: I1209 17:13:43.756118 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:13:44 crc kubenswrapper[4954]: I1209 17:13:44.545496 4954 scope.go:117] "RemoveContainer" containerID="a5d1719cf6f2890787807b62cb8048e90a86df344d1a2e7f1e476fbea320c809" Dec 09 17:13:44 crc kubenswrapper[4954]: I1209 17:13:44.567044 4954 scope.go:117] "RemoveContainer" containerID="687ca25af671c94a5c36d52a18645371434b84712dda6eefb6d80d456df6e1f3" Dec 09 17:13:44 crc kubenswrapper[4954]: I1209 17:13:44.599440 4954 scope.go:117] "RemoveContainer" containerID="dc87a87a32181d25699fbfe57266b887400734344545a9da5b532a99eedaca0c" Dec 09 17:14:13 crc kubenswrapper[4954]: I1209 17:14:13.755479 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:14:13 crc kubenswrapper[4954]: I1209 17:14:13.756105 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:14:13 crc kubenswrapper[4954]: I1209 17:14:13.756154 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:14:13 crc kubenswrapper[4954]: I1209 17:14:13.756808 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"809e10874828ead4756b40cdd641829a7300ce5c7ed7fd1b00389ceadbd9262b"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:14:13 crc kubenswrapper[4954]: I1209 17:14:13.756858 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://809e10874828ead4756b40cdd641829a7300ce5c7ed7fd1b00389ceadbd9262b" gracePeriod=600 Dec 09 17:14:14 crc kubenswrapper[4954]: I1209 17:14:14.094615 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="809e10874828ead4756b40cdd641829a7300ce5c7ed7fd1b00389ceadbd9262b" exitCode=0 Dec 09 17:14:14 crc kubenswrapper[4954]: I1209 17:14:14.094699 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"809e10874828ead4756b40cdd641829a7300ce5c7ed7fd1b00389ceadbd9262b"} Dec 09 17:14:14 crc kubenswrapper[4954]: I1209 17:14:14.094972 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"350d42655e86cea5ef6bd503b1c810df0baa263d9c78db2ee7b81665d25cfe98"} Dec 09 17:14:14 crc kubenswrapper[4954]: I1209 17:14:14.095005 4954 scope.go:117] "RemoveContainer" containerID="ea6d5b701a579f932d38d9e389879e05d1afbb8ef2ed598058ec1bc38f02d1f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.173497 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5"] Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.176448 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.180721 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.181105 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.192743 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5"] Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.201913 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-config-volume\") pod \"collect-profiles-29421675-cp6f5\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.201967 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-secret-volume\") pod \"collect-profiles-29421675-cp6f5\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.202101 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4hhv\" (UniqueName: \"kubernetes.io/projected/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-kube-api-access-t4hhv\") pod \"collect-profiles-29421675-cp6f5\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.303680 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-config-volume\") pod \"collect-profiles-29421675-cp6f5\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.303745 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-secret-volume\") pod \"collect-profiles-29421675-cp6f5\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.303833 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4hhv\" (UniqueName: \"kubernetes.io/projected/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-kube-api-access-t4hhv\") pod \"collect-profiles-29421675-cp6f5\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.305029 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-config-volume\") pod \"collect-profiles-29421675-cp6f5\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.312750 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-secret-volume\") pod \"collect-profiles-29421675-cp6f5\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.321452 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4hhv\" (UniqueName: \"kubernetes.io/projected/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-kube-api-access-t4hhv\") pod \"collect-profiles-29421675-cp6f5\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.501937 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:00 crc kubenswrapper[4954]: I1209 17:15:00.946733 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5"] Dec 09 17:15:01 crc kubenswrapper[4954]: I1209 17:15:01.417100 4954 generic.go:334] "Generic (PLEG): container finished" podID="26be4df9-9c9d-47f7-b334-bbffc33b9ccb" containerID="5d9cbe15162d8809d8457e6a60e98676eb62c0335eca7480f4a8746503912680" exitCode=0 Dec 09 17:15:01 crc kubenswrapper[4954]: I1209 17:15:01.417222 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" event={"ID":"26be4df9-9c9d-47f7-b334-bbffc33b9ccb","Type":"ContainerDied","Data":"5d9cbe15162d8809d8457e6a60e98676eb62c0335eca7480f4a8746503912680"} Dec 09 17:15:01 crc kubenswrapper[4954]: I1209 17:15:01.417460 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" event={"ID":"26be4df9-9c9d-47f7-b334-bbffc33b9ccb","Type":"ContainerStarted","Data":"04ff32516fc86bbdfefe870ffbf20f31002c547f624b7e158df46d085f83cfc4"} Dec 09 17:15:02 crc kubenswrapper[4954]: I1209 17:15:02.714264 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:02 crc kubenswrapper[4954]: I1209 17:15:02.853507 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4hhv\" (UniqueName: \"kubernetes.io/projected/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-kube-api-access-t4hhv\") pod \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " Dec 09 17:15:02 crc kubenswrapper[4954]: I1209 17:15:02.854153 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-config-volume\") pod \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " Dec 09 17:15:02 crc kubenswrapper[4954]: I1209 17:15:02.854269 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-secret-volume\") pod \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\" (UID: \"26be4df9-9c9d-47f7-b334-bbffc33b9ccb\") " Dec 09 17:15:02 crc kubenswrapper[4954]: I1209 17:15:02.855056 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-config-volume" (OuterVolumeSpecName: "config-volume") pod "26be4df9-9c9d-47f7-b334-bbffc33b9ccb" (UID: "26be4df9-9c9d-47f7-b334-bbffc33b9ccb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:15:02 crc kubenswrapper[4954]: I1209 17:15:02.855472 4954 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 17:15:02 crc kubenswrapper[4954]: I1209 17:15:02.864016 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-kube-api-access-t4hhv" (OuterVolumeSpecName: "kube-api-access-t4hhv") pod "26be4df9-9c9d-47f7-b334-bbffc33b9ccb" (UID: "26be4df9-9c9d-47f7-b334-bbffc33b9ccb"). InnerVolumeSpecName "kube-api-access-t4hhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:15:02 crc kubenswrapper[4954]: I1209 17:15:02.864876 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "26be4df9-9c9d-47f7-b334-bbffc33b9ccb" (UID: "26be4df9-9c9d-47f7-b334-bbffc33b9ccb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:15:02 crc kubenswrapper[4954]: I1209 17:15:02.956434 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4hhv\" (UniqueName: \"kubernetes.io/projected/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-kube-api-access-t4hhv\") on node \"crc\" DevicePath \"\"" Dec 09 17:15:02 crc kubenswrapper[4954]: I1209 17:15:02.956494 4954 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26be4df9-9c9d-47f7-b334-bbffc33b9ccb-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 17:15:03 crc kubenswrapper[4954]: I1209 17:15:03.437056 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" event={"ID":"26be4df9-9c9d-47f7-b334-bbffc33b9ccb","Type":"ContainerDied","Data":"04ff32516fc86bbdfefe870ffbf20f31002c547f624b7e158df46d085f83cfc4"} Dec 09 17:15:03 crc kubenswrapper[4954]: I1209 17:15:03.437132 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04ff32516fc86bbdfefe870ffbf20f31002c547f624b7e158df46d085f83cfc4" Dec 09 17:15:03 crc kubenswrapper[4954]: I1209 17:15:03.437537 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5" Dec 09 17:15:44 crc kubenswrapper[4954]: I1209 17:15:44.659627 4954 scope.go:117] "RemoveContainer" containerID="d8767a284227046fd48b1de5c8cb57dfbb846c01b63269b6c76666feef584ec6" Dec 09 17:15:44 crc kubenswrapper[4954]: I1209 17:15:44.684681 4954 scope.go:117] "RemoveContainer" containerID="cfc1421f4b77cb3cb6ab534ac714be0d504d6433c614a0290944aafe2e3d75a4" Dec 09 17:15:44 crc kubenswrapper[4954]: I1209 17:15:44.726548 4954 scope.go:117] "RemoveContainer" containerID="3404ff866c81781c3de706655cce7b77113eec6d7d280f18a431d3821b327c18" Dec 09 17:15:44 crc kubenswrapper[4954]: I1209 17:15:44.757069 4954 scope.go:117] "RemoveContainer" containerID="44b3a701ba31ae955f765ccd1957f59ceeb7e35d1f484b4529a86182192f8779" Dec 09 17:15:44 crc kubenswrapper[4954]: I1209 17:15:44.796653 4954 scope.go:117] "RemoveContainer" containerID="57edee372489592207069d713a70d1b12993c4c614266b349c936e438ad99523" Dec 09 17:15:44 crc kubenswrapper[4954]: I1209 17:15:44.829025 4954 scope.go:117] "RemoveContainer" containerID="a380e44d2794353a49b811e74d2a9168dc370fc2b3410331d23258074013c7cf" Dec 09 17:16:43 crc kubenswrapper[4954]: I1209 17:16:43.755921 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:16:43 crc kubenswrapper[4954]: I1209 17:16:43.757018 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:17:13 crc kubenswrapper[4954]: I1209 17:17:13.756281 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:17:13 crc kubenswrapper[4954]: I1209 17:17:13.757290 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:17:43 crc kubenswrapper[4954]: I1209 17:17:43.756741 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:17:43 crc kubenswrapper[4954]: I1209 17:17:43.757912 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:17:43 crc kubenswrapper[4954]: I1209 17:17:43.758010 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:17:43 crc kubenswrapper[4954]: I1209 17:17:43.759430 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"350d42655e86cea5ef6bd503b1c810df0baa263d9c78db2ee7b81665d25cfe98"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:17:43 crc kubenswrapper[4954]: I1209 17:17:43.759587 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://350d42655e86cea5ef6bd503b1c810df0baa263d9c78db2ee7b81665d25cfe98" gracePeriod=600 Dec 09 17:17:44 crc kubenswrapper[4954]: I1209 17:17:44.693088 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="350d42655e86cea5ef6bd503b1c810df0baa263d9c78db2ee7b81665d25cfe98" exitCode=0 Dec 09 17:17:44 crc kubenswrapper[4954]: I1209 17:17:44.693195 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"350d42655e86cea5ef6bd503b1c810df0baa263d9c78db2ee7b81665d25cfe98"} Dec 09 17:17:44 crc kubenswrapper[4954]: I1209 17:17:44.694093 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"4eb79bd8a7242b0656d839869821c630e99a7a1a245bd7247948b5ac5e84baa8"} Dec 09 17:17:44 crc kubenswrapper[4954]: I1209 17:17:44.694143 4954 scope.go:117] "RemoveContainer" containerID="809e10874828ead4756b40cdd641829a7300ce5c7ed7fd1b00389ceadbd9262b" Dec 09 17:20:13 crc kubenswrapper[4954]: I1209 17:20:13.755756 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:20:13 crc kubenswrapper[4954]: I1209 17:20:13.756387 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.059341 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv"] Dec 09 17:20:40 crc kubenswrapper[4954]: E1209 17:20:40.060460 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26be4df9-9c9d-47f7-b334-bbffc33b9ccb" containerName="collect-profiles" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.060477 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="26be4df9-9c9d-47f7-b334-bbffc33b9ccb" containerName="collect-profiles" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.060819 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="26be4df9-9c9d-47f7-b334-bbffc33b9ccb" containerName="collect-profiles" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.062968 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.067459 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.079794 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv"] Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.217487 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.217628 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rj55\" (UniqueName: \"kubernetes.io/projected/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-kube-api-access-9rj55\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.217673 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.319545 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.319672 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rj55\" (UniqueName: \"kubernetes.io/projected/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-kube-api-access-9rj55\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.319713 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.320144 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.320157 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.340359 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rj55\" (UniqueName: \"kubernetes.io/projected/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-kube-api-access-9rj55\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.388525 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.844954 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv"] Dec 09 17:20:40 crc kubenswrapper[4954]: I1209 17:20:40.905083 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" event={"ID":"a457e4b5-7bbb-480a-ae97-7cdfc2822dec","Type":"ContainerStarted","Data":"421fe7d2ab8e3baf71b47d98de40633c6567caad6cfc0d01da4108cf96ad3428"} Dec 09 17:20:41 crc kubenswrapper[4954]: I1209 17:20:41.796630 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z82cb"] Dec 09 17:20:41 crc kubenswrapper[4954]: I1209 17:20:41.798409 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:41 crc kubenswrapper[4954]: I1209 17:20:41.808710 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z82cb"] Dec 09 17:20:41 crc kubenswrapper[4954]: I1209 17:20:41.913363 4954 generic.go:334] "Generic (PLEG): container finished" podID="a457e4b5-7bbb-480a-ae97-7cdfc2822dec" containerID="4e98551ae4850c42388a03c6a438dbc8159be73f8aa0a168a5bbc757662a7051" exitCode=0 Dec 09 17:20:41 crc kubenswrapper[4954]: I1209 17:20:41.913412 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" event={"ID":"a457e4b5-7bbb-480a-ae97-7cdfc2822dec","Type":"ContainerDied","Data":"4e98551ae4850c42388a03c6a438dbc8159be73f8aa0a168a5bbc757662a7051"} Dec 09 17:20:41 crc kubenswrapper[4954]: I1209 17:20:41.915173 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 17:20:41 crc kubenswrapper[4954]: I1209 17:20:41.953547 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2glgz\" (UniqueName: \"kubernetes.io/projected/3da628af-03e1-482b-8e00-e7853292bc34-kube-api-access-2glgz\") pod \"redhat-operators-z82cb\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:41 crc kubenswrapper[4954]: I1209 17:20:41.953714 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-utilities\") pod \"redhat-operators-z82cb\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:41 crc kubenswrapper[4954]: I1209 17:20:41.953787 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-catalog-content\") pod \"redhat-operators-z82cb\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.055421 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-utilities\") pod \"redhat-operators-z82cb\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.055468 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-catalog-content\") pod \"redhat-operators-z82cb\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.055497 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2glgz\" (UniqueName: \"kubernetes.io/projected/3da628af-03e1-482b-8e00-e7853292bc34-kube-api-access-2glgz\") pod \"redhat-operators-z82cb\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.056080 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-catalog-content\") pod \"redhat-operators-z82cb\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.056101 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-utilities\") pod \"redhat-operators-z82cb\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.076666 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2glgz\" (UniqueName: \"kubernetes.io/projected/3da628af-03e1-482b-8e00-e7853292bc34-kube-api-access-2glgz\") pod \"redhat-operators-z82cb\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.115224 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.561372 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z82cb"] Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.921229 4954 generic.go:334] "Generic (PLEG): container finished" podID="3da628af-03e1-482b-8e00-e7853292bc34" containerID="cba25ac65b9f6251c7091ac8e2ac946f40e689ef94c95c1dd690c84d4236b4ce" exitCode=0 Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.921287 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z82cb" event={"ID":"3da628af-03e1-482b-8e00-e7853292bc34","Type":"ContainerDied","Data":"cba25ac65b9f6251c7091ac8e2ac946f40e689ef94c95c1dd690c84d4236b4ce"} Dec 09 17:20:42 crc kubenswrapper[4954]: I1209 17:20:42.921583 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z82cb" event={"ID":"3da628af-03e1-482b-8e00-e7853292bc34","Type":"ContainerStarted","Data":"6ea1b53f63e708a5369f038f9ad7af89915ab8b4b14ca0ad5eb41d1bbefa63c3"} Dec 09 17:20:43 crc kubenswrapper[4954]: I1209 17:20:43.755476 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:20:43 crc kubenswrapper[4954]: I1209 17:20:43.755609 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:20:43 crc kubenswrapper[4954]: I1209 17:20:43.933992 4954 generic.go:334] "Generic (PLEG): container finished" podID="a457e4b5-7bbb-480a-ae97-7cdfc2822dec" containerID="fcf9a668b29d641bd5db1365adbb7895b49e8d862969bb7a4dd738d153382e5b" exitCode=0 Dec 09 17:20:43 crc kubenswrapper[4954]: I1209 17:20:43.934114 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" event={"ID":"a457e4b5-7bbb-480a-ae97-7cdfc2822dec","Type":"ContainerDied","Data":"fcf9a668b29d641bd5db1365adbb7895b49e8d862969bb7a4dd738d153382e5b"} Dec 09 17:20:44 crc kubenswrapper[4954]: I1209 17:20:44.943232 4954 generic.go:334] "Generic (PLEG): container finished" podID="a457e4b5-7bbb-480a-ae97-7cdfc2822dec" containerID="c5aece78f1db989eab2cc0000ce79c8184bce89e82bde673eff52a9f9449e930" exitCode=0 Dec 09 17:20:44 crc kubenswrapper[4954]: I1209 17:20:44.943310 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" event={"ID":"a457e4b5-7bbb-480a-ae97-7cdfc2822dec","Type":"ContainerDied","Data":"c5aece78f1db989eab2cc0000ce79c8184bce89e82bde673eff52a9f9449e930"} Dec 09 17:20:44 crc kubenswrapper[4954]: I1209 17:20:44.945834 4954 generic.go:334] "Generic (PLEG): container finished" podID="3da628af-03e1-482b-8e00-e7853292bc34" containerID="896276a358a2410340f390c52a5f954d5bc91327c8661fab87d49e450a2d24b7" exitCode=0 Dec 09 17:20:44 crc kubenswrapper[4954]: I1209 17:20:44.945864 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z82cb" event={"ID":"3da628af-03e1-482b-8e00-e7853292bc34","Type":"ContainerDied","Data":"896276a358a2410340f390c52a5f954d5bc91327c8661fab87d49e450a2d24b7"} Dec 09 17:20:45 crc kubenswrapper[4954]: I1209 17:20:45.953873 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z82cb" event={"ID":"3da628af-03e1-482b-8e00-e7853292bc34","Type":"ContainerStarted","Data":"51efeb5bb06fb45528b06b68ffa892ec56d4e1363ed9ed07cc6eb515a4f5110e"} Dec 09 17:20:45 crc kubenswrapper[4954]: I1209 17:20:45.991865 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z82cb" podStartSLOduration=2.33497564 podStartE2EDuration="4.991829925s" podCreationTimestamp="2025-12-09 17:20:41 +0000 UTC" firstStartedPulling="2025-12-09 17:20:42.924071242 +0000 UTC m=+1439.312245062" lastFinishedPulling="2025-12-09 17:20:45.580925527 +0000 UTC m=+1441.969099347" observedRunningTime="2025-12-09 17:20:45.986328203 +0000 UTC m=+1442.374502043" watchObservedRunningTime="2025-12-09 17:20:45.991829925 +0000 UTC m=+1442.380003745" Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.379764 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.538336 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rj55\" (UniqueName: \"kubernetes.io/projected/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-kube-api-access-9rj55\") pod \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.538524 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-util\") pod \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.538713 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-bundle\") pod \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\" (UID: \"a457e4b5-7bbb-480a-ae97-7cdfc2822dec\") " Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.539308 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-bundle" (OuterVolumeSpecName: "bundle") pod "a457e4b5-7bbb-480a-ae97-7cdfc2822dec" (UID: "a457e4b5-7bbb-480a-ae97-7cdfc2822dec"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.546390 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-kube-api-access-9rj55" (OuterVolumeSpecName: "kube-api-access-9rj55") pod "a457e4b5-7bbb-480a-ae97-7cdfc2822dec" (UID: "a457e4b5-7bbb-480a-ae97-7cdfc2822dec"). InnerVolumeSpecName "kube-api-access-9rj55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.548264 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-util" (OuterVolumeSpecName: "util") pod "a457e4b5-7bbb-480a-ae97-7cdfc2822dec" (UID: "a457e4b5-7bbb-480a-ae97-7cdfc2822dec"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.640724 4954 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.640765 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rj55\" (UniqueName: \"kubernetes.io/projected/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-kube-api-access-9rj55\") on node \"crc\" DevicePath \"\"" Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.640778 4954 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a457e4b5-7bbb-480a-ae97-7cdfc2822dec-util\") on node \"crc\" DevicePath \"\"" Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.961669 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" event={"ID":"a457e4b5-7bbb-480a-ae97-7cdfc2822dec","Type":"ContainerDied","Data":"421fe7d2ab8e3baf71b47d98de40633c6567caad6cfc0d01da4108cf96ad3428"} Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.961736 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="421fe7d2ab8e3baf71b47d98de40633c6567caad6cfc0d01da4108cf96ad3428" Dec 09 17:20:46 crc kubenswrapper[4954]: I1209 17:20:46.961756 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.210103 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr"] Dec 09 17:20:51 crc kubenswrapper[4954]: E1209 17:20:51.210878 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a457e4b5-7bbb-480a-ae97-7cdfc2822dec" containerName="pull" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.210900 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a457e4b5-7bbb-480a-ae97-7cdfc2822dec" containerName="pull" Dec 09 17:20:51 crc kubenswrapper[4954]: E1209 17:20:51.210912 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a457e4b5-7bbb-480a-ae97-7cdfc2822dec" containerName="extract" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.210919 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a457e4b5-7bbb-480a-ae97-7cdfc2822dec" containerName="extract" Dec 09 17:20:51 crc kubenswrapper[4954]: E1209 17:20:51.210951 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a457e4b5-7bbb-480a-ae97-7cdfc2822dec" containerName="util" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.210959 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a457e4b5-7bbb-480a-ae97-7cdfc2822dec" containerName="util" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.211093 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a457e4b5-7bbb-480a-ae97-7cdfc2822dec" containerName="extract" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.211859 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.220132 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.222097 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.223062 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-5v58g" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.230093 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr"] Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.316118 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr8zg\" (UniqueName: \"kubernetes.io/projected/f365441f-c01b-4f49-834c-30326629074e-kube-api-access-cr8zg\") pod \"nmstate-operator-5b5b58f5c8-682nr\" (UID: \"f365441f-c01b-4f49-834c-30326629074e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.418046 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr8zg\" (UniqueName: \"kubernetes.io/projected/f365441f-c01b-4f49-834c-30326629074e-kube-api-access-cr8zg\") pod \"nmstate-operator-5b5b58f5c8-682nr\" (UID: \"f365441f-c01b-4f49-834c-30326629074e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.437716 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr8zg\" (UniqueName: \"kubernetes.io/projected/f365441f-c01b-4f49-834c-30326629074e-kube-api-access-cr8zg\") pod \"nmstate-operator-5b5b58f5c8-682nr\" (UID: \"f365441f-c01b-4f49-834c-30326629074e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr" Dec 09 17:20:51 crc kubenswrapper[4954]: I1209 17:20:51.542899 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr" Dec 09 17:20:52 crc kubenswrapper[4954]: I1209 17:20:52.022148 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr"] Dec 09 17:20:52 crc kubenswrapper[4954]: I1209 17:20:52.116360 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:52 crc kubenswrapper[4954]: I1209 17:20:52.116429 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:52 crc kubenswrapper[4954]: I1209 17:20:52.163669 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:53 crc kubenswrapper[4954]: I1209 17:20:53.003622 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr" event={"ID":"f365441f-c01b-4f49-834c-30326629074e","Type":"ContainerStarted","Data":"14c34912fb72ca5127f81ebd419d2b64074342573601bba011fe2f05f03735da"} Dec 09 17:20:53 crc kubenswrapper[4954]: I1209 17:20:53.046201 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:54 crc kubenswrapper[4954]: I1209 17:20:54.590372 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z82cb"] Dec 09 17:20:55 crc kubenswrapper[4954]: I1209 17:20:55.016041 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr" event={"ID":"f365441f-c01b-4f49-834c-30326629074e","Type":"ContainerStarted","Data":"bc50c3bd9de3217a56055a9d48168a1d6af99a682176ae6f8bb9cbe605b4fd54"} Dec 09 17:20:55 crc kubenswrapper[4954]: I1209 17:20:55.016254 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z82cb" podUID="3da628af-03e1-482b-8e00-e7853292bc34" containerName="registry-server" containerID="cri-o://51efeb5bb06fb45528b06b68ffa892ec56d4e1363ed9ed07cc6eb515a4f5110e" gracePeriod=2 Dec 09 17:20:55 crc kubenswrapper[4954]: I1209 17:20:55.038904 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-682nr" podStartSLOduration=1.532211401 podStartE2EDuration="4.038880698s" podCreationTimestamp="2025-12-09 17:20:51 +0000 UTC" firstStartedPulling="2025-12-09 17:20:52.032649176 +0000 UTC m=+1448.420822986" lastFinishedPulling="2025-12-09 17:20:54.539318443 +0000 UTC m=+1450.927492283" observedRunningTime="2025-12-09 17:20:55.033751938 +0000 UTC m=+1451.421925748" watchObservedRunningTime="2025-12-09 17:20:55.038880698 +0000 UTC m=+1451.427054518" Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.034693 4954 generic.go:334] "Generic (PLEG): container finished" podID="3da628af-03e1-482b-8e00-e7853292bc34" containerID="51efeb5bb06fb45528b06b68ffa892ec56d4e1363ed9ed07cc6eb515a4f5110e" exitCode=0 Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.034778 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z82cb" event={"ID":"3da628af-03e1-482b-8e00-e7853292bc34","Type":"ContainerDied","Data":"51efeb5bb06fb45528b06b68ffa892ec56d4e1363ed9ed07cc6eb515a4f5110e"} Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.286128 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.419228 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-utilities\") pod \"3da628af-03e1-482b-8e00-e7853292bc34\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.419345 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2glgz\" (UniqueName: \"kubernetes.io/projected/3da628af-03e1-482b-8e00-e7853292bc34-kube-api-access-2glgz\") pod \"3da628af-03e1-482b-8e00-e7853292bc34\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.419405 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-catalog-content\") pod \"3da628af-03e1-482b-8e00-e7853292bc34\" (UID: \"3da628af-03e1-482b-8e00-e7853292bc34\") " Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.420083 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-utilities" (OuterVolumeSpecName: "utilities") pod "3da628af-03e1-482b-8e00-e7853292bc34" (UID: "3da628af-03e1-482b-8e00-e7853292bc34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.425150 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3da628af-03e1-482b-8e00-e7853292bc34-kube-api-access-2glgz" (OuterVolumeSpecName: "kube-api-access-2glgz") pod "3da628af-03e1-482b-8e00-e7853292bc34" (UID: "3da628af-03e1-482b-8e00-e7853292bc34"). InnerVolumeSpecName "kube-api-access-2glgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.522128 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.522573 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2glgz\" (UniqueName: \"kubernetes.io/projected/3da628af-03e1-482b-8e00-e7853292bc34-kube-api-access-2glgz\") on node \"crc\" DevicePath \"\"" Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.523117 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3da628af-03e1-482b-8e00-e7853292bc34" (UID: "3da628af-03e1-482b-8e00-e7853292bc34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:20:57 crc kubenswrapper[4954]: I1209 17:20:57.624374 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3da628af-03e1-482b-8e00-e7853292bc34-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:20:58 crc kubenswrapper[4954]: I1209 17:20:58.046383 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z82cb" event={"ID":"3da628af-03e1-482b-8e00-e7853292bc34","Type":"ContainerDied","Data":"6ea1b53f63e708a5369f038f9ad7af89915ab8b4b14ca0ad5eb41d1bbefa63c3"} Dec 09 17:20:58 crc kubenswrapper[4954]: I1209 17:20:58.046472 4954 scope.go:117] "RemoveContainer" containerID="51efeb5bb06fb45528b06b68ffa892ec56d4e1363ed9ed07cc6eb515a4f5110e" Dec 09 17:20:58 crc kubenswrapper[4954]: I1209 17:20:58.046484 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z82cb" Dec 09 17:20:58 crc kubenswrapper[4954]: I1209 17:20:58.066150 4954 scope.go:117] "RemoveContainer" containerID="896276a358a2410340f390c52a5f954d5bc91327c8661fab87d49e450a2d24b7" Dec 09 17:20:58 crc kubenswrapper[4954]: I1209 17:20:58.087823 4954 scope.go:117] "RemoveContainer" containerID="cba25ac65b9f6251c7091ac8e2ac946f40e689ef94c95c1dd690c84d4236b4ce" Dec 09 17:20:58 crc kubenswrapper[4954]: I1209 17:20:58.098414 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z82cb"] Dec 09 17:20:58 crc kubenswrapper[4954]: I1209 17:20:58.104103 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z82cb"] Dec 09 17:20:58 crc kubenswrapper[4954]: I1209 17:20:58.131145 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3da628af-03e1-482b-8e00-e7853292bc34" path="/var/lib/kubelet/pods/3da628af-03e1-482b-8e00-e7853292bc34/volumes" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.027462 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm"] Dec 09 17:21:00 crc kubenswrapper[4954]: E1209 17:21:00.028193 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da628af-03e1-482b-8e00-e7853292bc34" containerName="registry-server" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.028210 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da628af-03e1-482b-8e00-e7853292bc34" containerName="registry-server" Dec 09 17:21:00 crc kubenswrapper[4954]: E1209 17:21:00.028230 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da628af-03e1-482b-8e00-e7853292bc34" containerName="extract-utilities" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.028239 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da628af-03e1-482b-8e00-e7853292bc34" containerName="extract-utilities" Dec 09 17:21:00 crc kubenswrapper[4954]: E1209 17:21:00.028251 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da628af-03e1-482b-8e00-e7853292bc34" containerName="extract-content" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.028259 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da628af-03e1-482b-8e00-e7853292bc34" containerName="extract-content" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.028403 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="3da628af-03e1-482b-8e00-e7853292bc34" containerName="registry-server" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.029362 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.033172 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-s8htp" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.033761 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q"] Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.035098 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.041652 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.052955 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q"] Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.063093 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm"] Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.068550 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-k85lr"] Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.069681 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.080733 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbb9b\" (UniqueName: \"kubernetes.io/projected/a3a905e4-153e-46de-879b-45e0870277b4-kube-api-access-xbb9b\") pod \"nmstate-metrics-7f946cbc9-r7xlm\" (UID: \"a3a905e4-153e-46de-879b-45e0870277b4\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.080817 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8b54c655-482f-4cfa-8029-c9fd0da67d17-ovs-socket\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.081007 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8b54c655-482f-4cfa-8029-c9fd0da67d17-dbus-socket\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.081089 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/71dff39f-cf61-434a-9c53-ad96e18d363f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-rls9q\" (UID: \"71dff39f-cf61-434a-9c53-ad96e18d363f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.081190 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8b54c655-482f-4cfa-8029-c9fd0da67d17-nmstate-lock\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.081211 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dl76\" (UniqueName: \"kubernetes.io/projected/8b54c655-482f-4cfa-8029-c9fd0da67d17-kube-api-access-5dl76\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.081260 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkzgv\" (UniqueName: \"kubernetes.io/projected/71dff39f-cf61-434a-9c53-ad96e18d363f-kube-api-access-nkzgv\") pod \"nmstate-webhook-5f6d4c5ccb-rls9q\" (UID: \"71dff39f-cf61-434a-9c53-ad96e18d363f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.182435 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/71dff39f-cf61-434a-9c53-ad96e18d363f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-rls9q\" (UID: \"71dff39f-cf61-434a-9c53-ad96e18d363f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.182487 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8b54c655-482f-4cfa-8029-c9fd0da67d17-nmstate-lock\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.182513 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dl76\" (UniqueName: \"kubernetes.io/projected/8b54c655-482f-4cfa-8029-c9fd0da67d17-kube-api-access-5dl76\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.182547 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkzgv\" (UniqueName: \"kubernetes.io/projected/71dff39f-cf61-434a-9c53-ad96e18d363f-kube-api-access-nkzgv\") pod \"nmstate-webhook-5f6d4c5ccb-rls9q\" (UID: \"71dff39f-cf61-434a-9c53-ad96e18d363f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.182623 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbb9b\" (UniqueName: \"kubernetes.io/projected/a3a905e4-153e-46de-879b-45e0870277b4-kube-api-access-xbb9b\") pod \"nmstate-metrics-7f946cbc9-r7xlm\" (UID: \"a3a905e4-153e-46de-879b-45e0870277b4\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.182680 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8b54c655-482f-4cfa-8029-c9fd0da67d17-ovs-socket\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.182712 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8b54c655-482f-4cfa-8029-c9fd0da67d17-nmstate-lock\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.182739 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8b54c655-482f-4cfa-8029-c9fd0da67d17-dbus-socket\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.182909 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8b54c655-482f-4cfa-8029-c9fd0da67d17-ovs-socket\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: E1209 17:21:00.182740 4954 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 09 17:21:00 crc kubenswrapper[4954]: E1209 17:21:00.183119 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/71dff39f-cf61-434a-9c53-ad96e18d363f-tls-key-pair podName:71dff39f-cf61-434a-9c53-ad96e18d363f nodeName:}" failed. No retries permitted until 2025-12-09 17:21:00.683084218 +0000 UTC m=+1457.071258218 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/71dff39f-cf61-434a-9c53-ad96e18d363f-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-rls9q" (UID: "71dff39f-cf61-434a-9c53-ad96e18d363f") : secret "openshift-nmstate-webhook" not found Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.183123 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8b54c655-482f-4cfa-8029-c9fd0da67d17-dbus-socket\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.206520 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd"] Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.207621 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.210979 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.211075 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.211148 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-szdgb" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.211351 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd"] Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.221544 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dl76\" (UniqueName: \"kubernetes.io/projected/8b54c655-482f-4cfa-8029-c9fd0da67d17-kube-api-access-5dl76\") pod \"nmstate-handler-k85lr\" (UID: \"8b54c655-482f-4cfa-8029-c9fd0da67d17\") " pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.221801 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkzgv\" (UniqueName: \"kubernetes.io/projected/71dff39f-cf61-434a-9c53-ad96e18d363f-kube-api-access-nkzgv\") pod \"nmstate-webhook-5f6d4c5ccb-rls9q\" (UID: \"71dff39f-cf61-434a-9c53-ad96e18d363f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.227050 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbb9b\" (UniqueName: \"kubernetes.io/projected/a3a905e4-153e-46de-879b-45e0870277b4-kube-api-access-xbb9b\") pod \"nmstate-metrics-7f946cbc9-r7xlm\" (UID: \"a3a905e4-153e-46de-879b-45e0870277b4\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.353378 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.385993 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/17ebcbf9-bb08-4a4f-b0c6-e081049376cf-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-tvljd\" (UID: \"17ebcbf9-bb08-4a4f-b0c6-e081049376cf\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.386561 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccgqs\" (UniqueName: \"kubernetes.io/projected/17ebcbf9-bb08-4a4f-b0c6-e081049376cf-kube-api-access-ccgqs\") pod \"nmstate-console-plugin-7fbb5f6569-tvljd\" (UID: \"17ebcbf9-bb08-4a4f-b0c6-e081049376cf\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.386612 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/17ebcbf9-bb08-4a4f-b0c6-e081049376cf-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-tvljd\" (UID: \"17ebcbf9-bb08-4a4f-b0c6-e081049376cf\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.390356 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.402506 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-bfd664845-r74kf"] Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.403398 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.433360 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-bfd664845-r74kf"] Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.488082 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-oauth-config\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.488168 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-service-ca\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.488205 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/17ebcbf9-bb08-4a4f-b0c6-e081049376cf-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-tvljd\" (UID: \"17ebcbf9-bb08-4a4f-b0c6-e081049376cf\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.488234 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrs6p\" (UniqueName: \"kubernetes.io/projected/bb473662-281d-46be-996e-02370393136b-kube-api-access-lrs6p\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.488268 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-oauth-serving-cert\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.488339 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccgqs\" (UniqueName: \"kubernetes.io/projected/17ebcbf9-bb08-4a4f-b0c6-e081049376cf-kube-api-access-ccgqs\") pod \"nmstate-console-plugin-7fbb5f6569-tvljd\" (UID: \"17ebcbf9-bb08-4a4f-b0c6-e081049376cf\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.488362 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-console-config\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.488420 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-trusted-ca-bundle\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.488448 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/17ebcbf9-bb08-4a4f-b0c6-e081049376cf-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-tvljd\" (UID: \"17ebcbf9-bb08-4a4f-b0c6-e081049376cf\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.488500 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-serving-cert\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.489932 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/17ebcbf9-bb08-4a4f-b0c6-e081049376cf-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-tvljd\" (UID: \"17ebcbf9-bb08-4a4f-b0c6-e081049376cf\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.495515 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/17ebcbf9-bb08-4a4f-b0c6-e081049376cf-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-tvljd\" (UID: \"17ebcbf9-bb08-4a4f-b0c6-e081049376cf\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.507453 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccgqs\" (UniqueName: \"kubernetes.io/projected/17ebcbf9-bb08-4a4f-b0c6-e081049376cf-kube-api-access-ccgqs\") pod \"nmstate-console-plugin-7fbb5f6569-tvljd\" (UID: \"17ebcbf9-bb08-4a4f-b0c6-e081049376cf\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.578178 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.591773 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-serving-cert\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.591850 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-oauth-config\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.591897 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-service-ca\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.591928 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrs6p\" (UniqueName: \"kubernetes.io/projected/bb473662-281d-46be-996e-02370393136b-kube-api-access-lrs6p\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.591955 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-oauth-serving-cert\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.592019 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-console-config\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.592047 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-trusted-ca-bundle\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.593194 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-service-ca\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.593750 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-oauth-serving-cert\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.594548 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-console-config\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.598736 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-trusted-ca-bundle\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.600449 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-serving-cert\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.607334 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-oauth-config\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.617625 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrs6p\" (UniqueName: \"kubernetes.io/projected/bb473662-281d-46be-996e-02370393136b-kube-api-access-lrs6p\") pod \"console-bfd664845-r74kf\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.697451 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/71dff39f-cf61-434a-9c53-ad96e18d363f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-rls9q\" (UID: \"71dff39f-cf61-434a-9c53-ad96e18d363f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.702190 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/71dff39f-cf61-434a-9c53-ad96e18d363f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-rls9q\" (UID: \"71dff39f-cf61-434a-9c53-ad96e18d363f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.774391 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm"] Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.774764 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.968337 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:00 crc kubenswrapper[4954]: I1209 17:21:00.997502 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd"] Dec 09 17:21:01 crc kubenswrapper[4954]: I1209 17:21:01.076880 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-bfd664845-r74kf"] Dec 09 17:21:01 crc kubenswrapper[4954]: W1209 17:21:01.098256 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb473662_281d_46be_996e_02370393136b.slice/crio-6bd945089dc93f02a3d79e9494098d989a604a78f2067273b6f6d3d26b955889 WatchSource:0}: Error finding container 6bd945089dc93f02a3d79e9494098d989a604a78f2067273b6f6d3d26b955889: Status 404 returned error can't find the container with id 6bd945089dc93f02a3d79e9494098d989a604a78f2067273b6f6d3d26b955889 Dec 09 17:21:01 crc kubenswrapper[4954]: I1209 17:21:01.104167 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm" event={"ID":"a3a905e4-153e-46de-879b-45e0870277b4","Type":"ContainerStarted","Data":"56bad88c1924170b52b9423b9de78622a4159da23dfcebdb3caa390e8105410a"} Dec 09 17:21:01 crc kubenswrapper[4954]: I1209 17:21:01.107482 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-k85lr" event={"ID":"8b54c655-482f-4cfa-8029-c9fd0da67d17","Type":"ContainerStarted","Data":"372076bac71961259540c32021bebb2b26d09591bc94739383deb66deef861f2"} Dec 09 17:21:01 crc kubenswrapper[4954]: I1209 17:21:01.109857 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" event={"ID":"17ebcbf9-bb08-4a4f-b0c6-e081049376cf","Type":"ContainerStarted","Data":"6825b6cfe7a8557da451bdcdbcd9b24bb3d29d53e3c3a8231874df226aacbe91"} Dec 09 17:21:01 crc kubenswrapper[4954]: I1209 17:21:01.437573 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q"] Dec 09 17:21:02 crc kubenswrapper[4954]: I1209 17:21:02.118751 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" event={"ID":"71dff39f-cf61-434a-9c53-ad96e18d363f","Type":"ContainerStarted","Data":"926da6fc2fe06d7cc499dee9edd7f6a9266f3b79efae178d8f0d905510e31c91"} Dec 09 17:21:02 crc kubenswrapper[4954]: I1209 17:21:02.132158 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-bfd664845-r74kf" event={"ID":"bb473662-281d-46be-996e-02370393136b","Type":"ContainerStarted","Data":"41cd1fb4c7e3b2a0b444d6593701517406c2247bb6e0d30d677685170ca77192"} Dec 09 17:21:02 crc kubenswrapper[4954]: I1209 17:21:02.132215 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-bfd664845-r74kf" event={"ID":"bb473662-281d-46be-996e-02370393136b","Type":"ContainerStarted","Data":"6bd945089dc93f02a3d79e9494098d989a604a78f2067273b6f6d3d26b955889"} Dec 09 17:21:02 crc kubenswrapper[4954]: I1209 17:21:02.139288 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-bfd664845-r74kf" podStartSLOduration=2.139271259 podStartE2EDuration="2.139271259s" podCreationTimestamp="2025-12-09 17:21:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:21:02.136458231 +0000 UTC m=+1458.524632051" watchObservedRunningTime="2025-12-09 17:21:02.139271259 +0000 UTC m=+1458.527445079" Dec 09 17:21:04 crc kubenswrapper[4954]: I1209 17:21:04.135202 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-k85lr" event={"ID":"8b54c655-482f-4cfa-8029-c9fd0da67d17","Type":"ContainerStarted","Data":"660e8153f09d84b21a5f2289e34691969aa441f2ea346428ed67d58de2b9629d"} Dec 09 17:21:04 crc kubenswrapper[4954]: I1209 17:21:04.136747 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" event={"ID":"71dff39f-cf61-434a-9c53-ad96e18d363f","Type":"ContainerStarted","Data":"fd7f6ec6d5deee3d10c346cf1b838d3c9367e0800fe2c0d57e8fed29f8ab6064"} Dec 09 17:21:04 crc kubenswrapper[4954]: I1209 17:21:04.136867 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:04 crc kubenswrapper[4954]: I1209 17:21:04.138092 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" event={"ID":"17ebcbf9-bb08-4a4f-b0c6-e081049376cf","Type":"ContainerStarted","Data":"83e37cde2c6383b98818942cc49899eeeab287faa8aae373153c39f94d6729bc"} Dec 09 17:21:04 crc kubenswrapper[4954]: I1209 17:21:04.139696 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm" event={"ID":"a3a905e4-153e-46de-879b-45e0870277b4","Type":"ContainerStarted","Data":"15ae0cf629a0212a46325c8a468bf9fa221f0dfce3a5a7946668f940d422cecc"} Dec 09 17:21:04 crc kubenswrapper[4954]: I1209 17:21:04.218269 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" podStartSLOduration=1.9013071099999999 podStartE2EDuration="4.218245934s" podCreationTimestamp="2025-12-09 17:21:00 +0000 UTC" firstStartedPulling="2025-12-09 17:21:01.444389305 +0000 UTC m=+1457.832563125" lastFinishedPulling="2025-12-09 17:21:03.761328109 +0000 UTC m=+1460.149501949" observedRunningTime="2025-12-09 17:21:04.213155494 +0000 UTC m=+1460.601329334" watchObservedRunningTime="2025-12-09 17:21:04.218245934 +0000 UTC m=+1460.606419754" Dec 09 17:21:05 crc kubenswrapper[4954]: I1209 17:21:05.170363 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tvljd" podStartSLOduration=2.406959946 podStartE2EDuration="5.170340878s" podCreationTimestamp="2025-12-09 17:21:00 +0000 UTC" firstStartedPulling="2025-12-09 17:21:01.01563742 +0000 UTC m=+1457.403811240" lastFinishedPulling="2025-12-09 17:21:03.779018352 +0000 UTC m=+1460.167192172" observedRunningTime="2025-12-09 17:21:05.163671329 +0000 UTC m=+1461.551845149" watchObservedRunningTime="2025-12-09 17:21:05.170340878 +0000 UTC m=+1461.558514698" Dec 09 17:21:05 crc kubenswrapper[4954]: I1209 17:21:05.192122 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-k85lr" podStartSLOduration=2.343559687 podStartE2EDuration="5.192104117s" podCreationTimestamp="2025-12-09 17:21:00 +0000 UTC" firstStartedPulling="2025-12-09 17:21:00.459831528 +0000 UTC m=+1456.848005338" lastFinishedPulling="2025-12-09 17:21:03.308375948 +0000 UTC m=+1459.696549768" observedRunningTime="2025-12-09 17:21:05.186313616 +0000 UTC m=+1461.574487446" watchObservedRunningTime="2025-12-09 17:21:05.192104117 +0000 UTC m=+1461.580277937" Dec 09 17:21:05 crc kubenswrapper[4954]: I1209 17:21:05.390668 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:07 crc kubenswrapper[4954]: I1209 17:21:07.173979 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm" event={"ID":"a3a905e4-153e-46de-879b-45e0870277b4","Type":"ContainerStarted","Data":"fdaef8c285eeacefd7d907e28af80dd2b1ebac39b5e7524b01f8fd970056b84f"} Dec 09 17:21:07 crc kubenswrapper[4954]: I1209 17:21:07.198886 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-r7xlm" podStartSLOduration=1.262478426 podStartE2EDuration="7.198867247s" podCreationTimestamp="2025-12-09 17:21:00 +0000 UTC" firstStartedPulling="2025-12-09 17:21:00.810776374 +0000 UTC m=+1457.198950194" lastFinishedPulling="2025-12-09 17:21:06.747165195 +0000 UTC m=+1463.135339015" observedRunningTime="2025-12-09 17:21:07.194319695 +0000 UTC m=+1463.582493515" watchObservedRunningTime="2025-12-09 17:21:07.198867247 +0000 UTC m=+1463.587041077" Dec 09 17:21:10 crc kubenswrapper[4954]: I1209 17:21:10.418298 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-k85lr" Dec 09 17:21:10 crc kubenswrapper[4954]: I1209 17:21:10.775065 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:10 crc kubenswrapper[4954]: I1209 17:21:10.775142 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:10 crc kubenswrapper[4954]: I1209 17:21:10.780501 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:11 crc kubenswrapper[4954]: I1209 17:21:11.207900 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:21:11 crc kubenswrapper[4954]: I1209 17:21:11.272951 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-v4pm7"] Dec 09 17:21:13 crc kubenswrapper[4954]: I1209 17:21:13.756221 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:21:13 crc kubenswrapper[4954]: I1209 17:21:13.756835 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:21:13 crc kubenswrapper[4954]: I1209 17:21:13.756885 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:21:13 crc kubenswrapper[4954]: I1209 17:21:13.757841 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4eb79bd8a7242b0656d839869821c630e99a7a1a245bd7247948b5ac5e84baa8"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:21:13 crc kubenswrapper[4954]: I1209 17:21:13.757905 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://4eb79bd8a7242b0656d839869821c630e99a7a1a245bd7247948b5ac5e84baa8" gracePeriod=600 Dec 09 17:21:14 crc kubenswrapper[4954]: I1209 17:21:14.225983 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="4eb79bd8a7242b0656d839869821c630e99a7a1a245bd7247948b5ac5e84baa8" exitCode=0 Dec 09 17:21:14 crc kubenswrapper[4954]: I1209 17:21:14.226389 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"4eb79bd8a7242b0656d839869821c630e99a7a1a245bd7247948b5ac5e84baa8"} Dec 09 17:21:14 crc kubenswrapper[4954]: I1209 17:21:14.226418 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d"} Dec 09 17:21:14 crc kubenswrapper[4954]: I1209 17:21:14.226433 4954 scope.go:117] "RemoveContainer" containerID="350d42655e86cea5ef6bd503b1c810df0baa263d9c78db2ee7b81665d25cfe98" Dec 09 17:21:20 crc kubenswrapper[4954]: I1209 17:21:20.975061 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-rls9q" Dec 09 17:21:34 crc kubenswrapper[4954]: I1209 17:21:34.884322 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m"] Dec 09 17:21:34 crc kubenswrapper[4954]: I1209 17:21:34.886389 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:34 crc kubenswrapper[4954]: I1209 17:21:34.888452 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 17:21:34 crc kubenswrapper[4954]: I1209 17:21:34.932475 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m"] Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.007306 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.007538 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nrfp\" (UniqueName: \"kubernetes.io/projected/97ccf7f7-c707-4236-bb60-ac38713c8793-kube-api-access-6nrfp\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.007719 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.109357 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.109442 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nrfp\" (UniqueName: \"kubernetes.io/projected/97ccf7f7-c707-4236-bb60-ac38713c8793-kube-api-access-6nrfp\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.109498 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.109967 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.110092 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.133113 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nrfp\" (UniqueName: \"kubernetes.io/projected/97ccf7f7-c707-4236-bb60-ac38713c8793-kube-api-access-6nrfp\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.203063 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:35 crc kubenswrapper[4954]: I1209 17:21:35.634366 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m"] Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.323011 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-v4pm7" podUID="7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" containerName="console" containerID="cri-o://244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38" gracePeriod=15 Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.388263 4954 generic.go:334] "Generic (PLEG): container finished" podID="97ccf7f7-c707-4236-bb60-ac38713c8793" containerID="509d4714b72abe29fc069091eabec0c4ef8f1bf3e2675cbcf0309eea59a90623" exitCode=0 Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.388305 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" event={"ID":"97ccf7f7-c707-4236-bb60-ac38713c8793","Type":"ContainerDied","Data":"509d4714b72abe29fc069091eabec0c4ef8f1bf3e2675cbcf0309eea59a90623"} Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.388333 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" event={"ID":"97ccf7f7-c707-4236-bb60-ac38713c8793","Type":"ContainerStarted","Data":"6250d8a8f15879fafc7fcbfcb4b3cfc1ae7942f92ffdd48fe827f45f3f836bb7"} Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.702083 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-v4pm7_7bb5ad5d-2235-47f8-8100-8b79d0d44fb0/console/0.log" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.702422 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.844544 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-oauth-serving-cert\") pod \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.844630 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9wtr\" (UniqueName: \"kubernetes.io/projected/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-kube-api-access-d9wtr\") pod \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.844673 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-serving-cert\") pod \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.844728 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-config\") pod \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.844791 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-trusted-ca-bundle\") pod \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.844832 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-oauth-config\") pod \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.844850 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-service-ca\") pod \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\" (UID: \"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0\") " Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.845294 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" (UID: "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.845816 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-service-ca" (OuterVolumeSpecName: "service-ca") pod "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" (UID: "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.845946 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-config" (OuterVolumeSpecName: "console-config") pod "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" (UID: "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.846219 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" (UID: "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.857886 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" (UID: "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.858241 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-kube-api-access-d9wtr" (OuterVolumeSpecName: "kube-api-access-d9wtr") pod "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" (UID: "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0"). InnerVolumeSpecName "kube-api-access-d9wtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.858453 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" (UID: "7bb5ad5d-2235-47f8-8100-8b79d0d44fb0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.948075 4954 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.948105 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9wtr\" (UniqueName: \"kubernetes.io/projected/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-kube-api-access-d9wtr\") on node \"crc\" DevicePath \"\"" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.948117 4954 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.948126 4954 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.948135 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.948143 4954 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:21:36 crc kubenswrapper[4954]: I1209 17:21:36.948153 4954 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 17:21:37 crc kubenswrapper[4954]: I1209 17:21:37.395303 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-v4pm7_7bb5ad5d-2235-47f8-8100-8b79d0d44fb0/console/0.log" Dec 09 17:21:37 crc kubenswrapper[4954]: I1209 17:21:37.395351 4954 generic.go:334] "Generic (PLEG): container finished" podID="7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" containerID="244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38" exitCode=2 Dec 09 17:21:37 crc kubenswrapper[4954]: I1209 17:21:37.395391 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-v4pm7" event={"ID":"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0","Type":"ContainerDied","Data":"244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38"} Dec 09 17:21:37 crc kubenswrapper[4954]: I1209 17:21:37.395424 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-v4pm7" event={"ID":"7bb5ad5d-2235-47f8-8100-8b79d0d44fb0","Type":"ContainerDied","Data":"317339b18d0b617eab6ed4f455e51414fe7e42cf6a08bd12f2606b411b42df6b"} Dec 09 17:21:37 crc kubenswrapper[4954]: I1209 17:21:37.395442 4954 scope.go:117] "RemoveContainer" containerID="244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38" Dec 09 17:21:37 crc kubenswrapper[4954]: I1209 17:21:37.395563 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-v4pm7" Dec 09 17:21:37 crc kubenswrapper[4954]: I1209 17:21:37.414071 4954 scope.go:117] "RemoveContainer" containerID="244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38" Dec 09 17:21:37 crc kubenswrapper[4954]: E1209 17:21:37.414713 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38\": container with ID starting with 244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38 not found: ID does not exist" containerID="244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38" Dec 09 17:21:37 crc kubenswrapper[4954]: I1209 17:21:37.414785 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38"} err="failed to get container status \"244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38\": rpc error: code = NotFound desc = could not find container \"244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38\": container with ID starting with 244a24fd25743340d62403950f53eff75ef94667a243d1fe51efd695bc72aa38 not found: ID does not exist" Dec 09 17:21:37 crc kubenswrapper[4954]: I1209 17:21:37.425936 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-v4pm7"] Dec 09 17:21:37 crc kubenswrapper[4954]: I1209 17:21:37.433421 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-v4pm7"] Dec 09 17:21:38 crc kubenswrapper[4954]: I1209 17:21:38.131330 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" path="/var/lib/kubelet/pods/7bb5ad5d-2235-47f8-8100-8b79d0d44fb0/volumes" Dec 09 17:21:38 crc kubenswrapper[4954]: I1209 17:21:38.406044 4954 generic.go:334] "Generic (PLEG): container finished" podID="97ccf7f7-c707-4236-bb60-ac38713c8793" containerID="c501e0fbfaa0f9e464ebfd7385e055fb3f0834aaf7be17f657734e223952d200" exitCode=0 Dec 09 17:21:38 crc kubenswrapper[4954]: I1209 17:21:38.406100 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" event={"ID":"97ccf7f7-c707-4236-bb60-ac38713c8793","Type":"ContainerDied","Data":"c501e0fbfaa0f9e464ebfd7385e055fb3f0834aaf7be17f657734e223952d200"} Dec 09 17:21:39 crc kubenswrapper[4954]: I1209 17:21:39.413564 4954 generic.go:334] "Generic (PLEG): container finished" podID="97ccf7f7-c707-4236-bb60-ac38713c8793" containerID="a6607ad6c5fa62e4506610b007c6d80307fbb3aae5ed0ab030a6df04e5fef25e" exitCode=0 Dec 09 17:21:39 crc kubenswrapper[4954]: I1209 17:21:39.413639 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" event={"ID":"97ccf7f7-c707-4236-bb60-ac38713c8793","Type":"ContainerDied","Data":"a6607ad6c5fa62e4506610b007c6d80307fbb3aae5ed0ab030a6df04e5fef25e"} Dec 09 17:21:40 crc kubenswrapper[4954]: I1209 17:21:40.674550 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:40 crc kubenswrapper[4954]: I1209 17:21:40.805756 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-util\") pod \"97ccf7f7-c707-4236-bb60-ac38713c8793\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " Dec 09 17:21:40 crc kubenswrapper[4954]: I1209 17:21:40.805843 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nrfp\" (UniqueName: \"kubernetes.io/projected/97ccf7f7-c707-4236-bb60-ac38713c8793-kube-api-access-6nrfp\") pod \"97ccf7f7-c707-4236-bb60-ac38713c8793\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " Dec 09 17:21:40 crc kubenswrapper[4954]: I1209 17:21:40.805897 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-bundle\") pod \"97ccf7f7-c707-4236-bb60-ac38713c8793\" (UID: \"97ccf7f7-c707-4236-bb60-ac38713c8793\") " Dec 09 17:21:40 crc kubenswrapper[4954]: I1209 17:21:40.807090 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-bundle" (OuterVolumeSpecName: "bundle") pod "97ccf7f7-c707-4236-bb60-ac38713c8793" (UID: "97ccf7f7-c707-4236-bb60-ac38713c8793"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:21:40 crc kubenswrapper[4954]: I1209 17:21:40.813771 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97ccf7f7-c707-4236-bb60-ac38713c8793-kube-api-access-6nrfp" (OuterVolumeSpecName: "kube-api-access-6nrfp") pod "97ccf7f7-c707-4236-bb60-ac38713c8793" (UID: "97ccf7f7-c707-4236-bb60-ac38713c8793"). InnerVolumeSpecName "kube-api-access-6nrfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:21:40 crc kubenswrapper[4954]: I1209 17:21:40.819432 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-util" (OuterVolumeSpecName: "util") pod "97ccf7f7-c707-4236-bb60-ac38713c8793" (UID: "97ccf7f7-c707-4236-bb60-ac38713c8793"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:21:40 crc kubenswrapper[4954]: I1209 17:21:40.907126 4954 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:21:40 crc kubenswrapper[4954]: I1209 17:21:40.907162 4954 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/97ccf7f7-c707-4236-bb60-ac38713c8793-util\") on node \"crc\" DevicePath \"\"" Dec 09 17:21:40 crc kubenswrapper[4954]: I1209 17:21:40.907171 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nrfp\" (UniqueName: \"kubernetes.io/projected/97ccf7f7-c707-4236-bb60-ac38713c8793-kube-api-access-6nrfp\") on node \"crc\" DevicePath \"\"" Dec 09 17:21:41 crc kubenswrapper[4954]: I1209 17:21:41.428294 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" event={"ID":"97ccf7f7-c707-4236-bb60-ac38713c8793","Type":"ContainerDied","Data":"6250d8a8f15879fafc7fcbfcb4b3cfc1ae7942f92ffdd48fe827f45f3f836bb7"} Dec 09 17:21:41 crc kubenswrapper[4954]: I1209 17:21:41.428615 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6250d8a8f15879fafc7fcbfcb4b3cfc1ae7942f92ffdd48fe827f45f3f836bb7" Dec 09 17:21:41 crc kubenswrapper[4954]: I1209 17:21:41.428365 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.374027 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-98gkb"] Dec 09 17:21:48 crc kubenswrapper[4954]: E1209 17:21:48.374831 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97ccf7f7-c707-4236-bb60-ac38713c8793" containerName="pull" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.374843 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="97ccf7f7-c707-4236-bb60-ac38713c8793" containerName="pull" Dec 09 17:21:48 crc kubenswrapper[4954]: E1209 17:21:48.374863 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" containerName="console" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.374869 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" containerName="console" Dec 09 17:21:48 crc kubenswrapper[4954]: E1209 17:21:48.374876 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97ccf7f7-c707-4236-bb60-ac38713c8793" containerName="util" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.374883 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="97ccf7f7-c707-4236-bb60-ac38713c8793" containerName="util" Dec 09 17:21:48 crc kubenswrapper[4954]: E1209 17:21:48.374893 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97ccf7f7-c707-4236-bb60-ac38713c8793" containerName="extract" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.374899 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="97ccf7f7-c707-4236-bb60-ac38713c8793" containerName="extract" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.375013 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bb5ad5d-2235-47f8-8100-8b79d0d44fb0" containerName="console" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.375029 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="97ccf7f7-c707-4236-bb60-ac38713c8793" containerName="extract" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.375905 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.389705 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-98gkb"] Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.528361 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rmmh\" (UniqueName: \"kubernetes.io/projected/00f163ab-337a-49bf-b3db-ca0c79496c1f-kube-api-access-6rmmh\") pod \"redhat-marketplace-98gkb\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.528816 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-catalog-content\") pod \"redhat-marketplace-98gkb\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.528886 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-utilities\") pod \"redhat-marketplace-98gkb\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.630160 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-utilities\") pod \"redhat-marketplace-98gkb\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.630322 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rmmh\" (UniqueName: \"kubernetes.io/projected/00f163ab-337a-49bf-b3db-ca0c79496c1f-kube-api-access-6rmmh\") pod \"redhat-marketplace-98gkb\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.630424 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-catalog-content\") pod \"redhat-marketplace-98gkb\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.630755 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-utilities\") pod \"redhat-marketplace-98gkb\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.631001 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-catalog-content\") pod \"redhat-marketplace-98gkb\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.652205 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rmmh\" (UniqueName: \"kubernetes.io/projected/00f163ab-337a-49bf-b3db-ca0c79496c1f-kube-api-access-6rmmh\") pod \"redhat-marketplace-98gkb\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:48 crc kubenswrapper[4954]: I1209 17:21:48.693754 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.152243 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-98gkb"] Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.477586 4954 generic.go:334] "Generic (PLEG): container finished" podID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerID="8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b" exitCode=0 Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.477673 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98gkb" event={"ID":"00f163ab-337a-49bf-b3db-ca0c79496c1f","Type":"ContainerDied","Data":"8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b"} Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.477709 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98gkb" event={"ID":"00f163ab-337a-49bf-b3db-ca0c79496c1f","Type":"ContainerStarted","Data":"dfb31cda9cfcc8e87335abab5743a70386bf7e7c9cc9cd4404723443f1c00817"} Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.903909 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8"] Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.905117 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.910350 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.910891 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-hbxqd" Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.911080 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.911323 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.911475 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 09 17:21:49 crc kubenswrapper[4954]: I1209 17:21:49.929987 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8"] Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.054237 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2-webhook-cert\") pod \"metallb-operator-controller-manager-7fb7cc7845-b4jd8\" (UID: \"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2\") " pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.054310 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2-apiservice-cert\") pod \"metallb-operator-controller-manager-7fb7cc7845-b4jd8\" (UID: \"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2\") " pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.054556 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j72f9\" (UniqueName: \"kubernetes.io/projected/4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2-kube-api-access-j72f9\") pod \"metallb-operator-controller-manager-7fb7cc7845-b4jd8\" (UID: \"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2\") " pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.156578 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2-webhook-cert\") pod \"metallb-operator-controller-manager-7fb7cc7845-b4jd8\" (UID: \"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2\") " pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.156690 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2-apiservice-cert\") pod \"metallb-operator-controller-manager-7fb7cc7845-b4jd8\" (UID: \"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2\") " pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.156773 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j72f9\" (UniqueName: \"kubernetes.io/projected/4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2-kube-api-access-j72f9\") pod \"metallb-operator-controller-manager-7fb7cc7845-b4jd8\" (UID: \"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2\") " pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.166459 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2-webhook-cert\") pod \"metallb-operator-controller-manager-7fb7cc7845-b4jd8\" (UID: \"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2\") " pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.182334 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2-apiservice-cert\") pod \"metallb-operator-controller-manager-7fb7cc7845-b4jd8\" (UID: \"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2\") " pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.195706 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j72f9\" (UniqueName: \"kubernetes.io/projected/4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2-kube-api-access-j72f9\") pod \"metallb-operator-controller-manager-7fb7cc7845-b4jd8\" (UID: \"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2\") " pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.227650 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.512496 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78"] Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.513969 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.524487 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.524785 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-g4b5p" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.524806 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.529446 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78"] Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.670027 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7-apiservice-cert\") pod \"metallb-operator-webhook-server-6b9d7ddc4d-l8z78\" (UID: \"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7\") " pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.670110 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7-webhook-cert\") pod \"metallb-operator-webhook-server-6b9d7ddc4d-l8z78\" (UID: \"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7\") " pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.670327 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-599gt\" (UniqueName: \"kubernetes.io/projected/f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7-kube-api-access-599gt\") pod \"metallb-operator-webhook-server-6b9d7ddc4d-l8z78\" (UID: \"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7\") " pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.772435 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-599gt\" (UniqueName: \"kubernetes.io/projected/f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7-kube-api-access-599gt\") pod \"metallb-operator-webhook-server-6b9d7ddc4d-l8z78\" (UID: \"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7\") " pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.773348 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7-apiservice-cert\") pod \"metallb-operator-webhook-server-6b9d7ddc4d-l8z78\" (UID: \"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7\") " pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.774442 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7-webhook-cert\") pod \"metallb-operator-webhook-server-6b9d7ddc4d-l8z78\" (UID: \"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7\") " pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.781258 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7-webhook-cert\") pod \"metallb-operator-webhook-server-6b9d7ddc4d-l8z78\" (UID: \"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7\") " pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.782835 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7-apiservice-cert\") pod \"metallb-operator-webhook-server-6b9d7ddc4d-l8z78\" (UID: \"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7\") " pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.797140 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-599gt\" (UniqueName: \"kubernetes.io/projected/f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7-kube-api-access-599gt\") pod \"metallb-operator-webhook-server-6b9d7ddc4d-l8z78\" (UID: \"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7\") " pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.801227 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8"] Dec 09 17:21:50 crc kubenswrapper[4954]: W1209 17:21:50.808973 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d80ef44_a45a_4e4f_b61f_d2b30d9c1fc2.slice/crio-acca71b88db57c6f8b379f475d38cb69ecc11d89c2651c7d3c7b22f26fb75809 WatchSource:0}: Error finding container acca71b88db57c6f8b379f475d38cb69ecc11d89c2651c7d3c7b22f26fb75809: Status 404 returned error can't find the container with id acca71b88db57c6f8b379f475d38cb69ecc11d89c2651c7d3c7b22f26fb75809 Dec 09 17:21:50 crc kubenswrapper[4954]: I1209 17:21:50.841638 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:51 crc kubenswrapper[4954]: I1209 17:21:51.296232 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78"] Dec 09 17:21:51 crc kubenswrapper[4954]: W1209 17:21:51.305060 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8aed5f9_f76a_4a19_92ae_0b11a2a3fee7.slice/crio-2d2c449947e96c7c63981d2d1ebd1f938bb7402a8fe4e5397c4ebd62b05d4523 WatchSource:0}: Error finding container 2d2c449947e96c7c63981d2d1ebd1f938bb7402a8fe4e5397c4ebd62b05d4523: Status 404 returned error can't find the container with id 2d2c449947e96c7c63981d2d1ebd1f938bb7402a8fe4e5397c4ebd62b05d4523 Dec 09 17:21:51 crc kubenswrapper[4954]: I1209 17:21:51.498073 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" event={"ID":"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7","Type":"ContainerStarted","Data":"2d2c449947e96c7c63981d2d1ebd1f938bb7402a8fe4e5397c4ebd62b05d4523"} Dec 09 17:21:51 crc kubenswrapper[4954]: I1209 17:21:51.499408 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" event={"ID":"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2","Type":"ContainerStarted","Data":"acca71b88db57c6f8b379f475d38cb69ecc11d89c2651c7d3c7b22f26fb75809"} Dec 09 17:21:51 crc kubenswrapper[4954]: I1209 17:21:51.502283 4954 generic.go:334] "Generic (PLEG): container finished" podID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerID="48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638" exitCode=0 Dec 09 17:21:51 crc kubenswrapper[4954]: I1209 17:21:51.502330 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98gkb" event={"ID":"00f163ab-337a-49bf-b3db-ca0c79496c1f","Type":"ContainerDied","Data":"48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638"} Dec 09 17:21:53 crc kubenswrapper[4954]: I1209 17:21:53.518868 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98gkb" event={"ID":"00f163ab-337a-49bf-b3db-ca0c79496c1f","Type":"ContainerStarted","Data":"9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1"} Dec 09 17:21:53 crc kubenswrapper[4954]: I1209 17:21:53.546155 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-98gkb" podStartSLOduration=2.329228589 podStartE2EDuration="5.546136999s" podCreationTimestamp="2025-12-09 17:21:48 +0000 UTC" firstStartedPulling="2025-12-09 17:21:49.479879392 +0000 UTC m=+1505.868053212" lastFinishedPulling="2025-12-09 17:21:52.696787802 +0000 UTC m=+1509.084961622" observedRunningTime="2025-12-09 17:21:53.543740174 +0000 UTC m=+1509.931914004" watchObservedRunningTime="2025-12-09 17:21:53.546136999 +0000 UTC m=+1509.934310819" Dec 09 17:21:58 crc kubenswrapper[4954]: I1209 17:21:58.557873 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" event={"ID":"f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7","Type":"ContainerStarted","Data":"4e209953ede9de6b112505d91d41ae784e8f23320857f8ffec9d40f8ef64c09a"} Dec 09 17:21:58 crc kubenswrapper[4954]: I1209 17:21:58.558665 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:21:58 crc kubenswrapper[4954]: I1209 17:21:58.561218 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" event={"ID":"4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2","Type":"ContainerStarted","Data":"cd9380414034d44a901aa581e59ec3c8f2d84277c232d8575f568ffe96c68541"} Dec 09 17:21:58 crc kubenswrapper[4954]: I1209 17:21:58.561373 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:21:58 crc kubenswrapper[4954]: I1209 17:21:58.575056 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" podStartSLOduration=2.484001591 podStartE2EDuration="8.575038789s" podCreationTimestamp="2025-12-09 17:21:50 +0000 UTC" firstStartedPulling="2025-12-09 17:21:51.308253893 +0000 UTC m=+1507.696427713" lastFinishedPulling="2025-12-09 17:21:57.399291091 +0000 UTC m=+1513.787464911" observedRunningTime="2025-12-09 17:21:58.573928203 +0000 UTC m=+1514.962102033" watchObservedRunningTime="2025-12-09 17:21:58.575038789 +0000 UTC m=+1514.963212609" Dec 09 17:21:58 crc kubenswrapper[4954]: I1209 17:21:58.601873 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" podStartSLOduration=3.041073251 podStartE2EDuration="9.601853885s" podCreationTimestamp="2025-12-09 17:21:49 +0000 UTC" firstStartedPulling="2025-12-09 17:21:50.814421176 +0000 UTC m=+1507.202594996" lastFinishedPulling="2025-12-09 17:21:57.37520181 +0000 UTC m=+1513.763375630" observedRunningTime="2025-12-09 17:21:58.599472741 +0000 UTC m=+1514.987646571" watchObservedRunningTime="2025-12-09 17:21:58.601853885 +0000 UTC m=+1514.990027705" Dec 09 17:21:58 crc kubenswrapper[4954]: I1209 17:21:58.693838 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:58 crc kubenswrapper[4954]: I1209 17:21:58.693890 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:58 crc kubenswrapper[4954]: I1209 17:21:58.752021 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:59 crc kubenswrapper[4954]: I1209 17:21:59.613907 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:21:59 crc kubenswrapper[4954]: I1209 17:21:59.665723 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-98gkb"] Dec 09 17:22:01 crc kubenswrapper[4954]: I1209 17:22:01.582810 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-98gkb" podUID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerName="registry-server" containerID="cri-o://9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1" gracePeriod=2 Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.530644 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.598133 4954 generic.go:334] "Generic (PLEG): container finished" podID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerID="9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1" exitCode=0 Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.598181 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98gkb" event={"ID":"00f163ab-337a-49bf-b3db-ca0c79496c1f","Type":"ContainerDied","Data":"9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1"} Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.598223 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98gkb" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.598252 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98gkb" event={"ID":"00f163ab-337a-49bf-b3db-ca0c79496c1f","Type":"ContainerDied","Data":"dfb31cda9cfcc8e87335abab5743a70386bf7e7c9cc9cd4404723443f1c00817"} Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.598279 4954 scope.go:117] "RemoveContainer" containerID="9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.621517 4954 scope.go:117] "RemoveContainer" containerID="48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.645716 4954 scope.go:117] "RemoveContainer" containerID="8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.666820 4954 scope.go:117] "RemoveContainer" containerID="9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1" Dec 09 17:22:02 crc kubenswrapper[4954]: E1209 17:22:02.667437 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1\": container with ID starting with 9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1 not found: ID does not exist" containerID="9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.667494 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1"} err="failed to get container status \"9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1\": rpc error: code = NotFound desc = could not find container \"9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1\": container with ID starting with 9ed5b47a3e561715dabe5d3a19b8d21ff888761a2e3cedc964be8ec307d6dbd1 not found: ID does not exist" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.667522 4954 scope.go:117] "RemoveContainer" containerID="48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638" Dec 09 17:22:02 crc kubenswrapper[4954]: E1209 17:22:02.667941 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638\": container with ID starting with 48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638 not found: ID does not exist" containerID="48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.667995 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638"} err="failed to get container status \"48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638\": rpc error: code = NotFound desc = could not find container \"48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638\": container with ID starting with 48b125befdc6f016d5158bb5ab76f80216f771a202e2bfe97e660a723cb91638 not found: ID does not exist" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.668050 4954 scope.go:117] "RemoveContainer" containerID="8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b" Dec 09 17:22:02 crc kubenswrapper[4954]: E1209 17:22:02.668441 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b\": container with ID starting with 8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b not found: ID does not exist" containerID="8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.668471 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b"} err="failed to get container status \"8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b\": rpc error: code = NotFound desc = could not find container \"8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b\": container with ID starting with 8925d3194addfd07fe91b3f09c29d62d7110317a951042da4480cb56750e107b not found: ID does not exist" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.673066 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rmmh\" (UniqueName: \"kubernetes.io/projected/00f163ab-337a-49bf-b3db-ca0c79496c1f-kube-api-access-6rmmh\") pod \"00f163ab-337a-49bf-b3db-ca0c79496c1f\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.673156 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-catalog-content\") pod \"00f163ab-337a-49bf-b3db-ca0c79496c1f\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.679401 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00f163ab-337a-49bf-b3db-ca0c79496c1f-kube-api-access-6rmmh" (OuterVolumeSpecName: "kube-api-access-6rmmh") pod "00f163ab-337a-49bf-b3db-ca0c79496c1f" (UID: "00f163ab-337a-49bf-b3db-ca0c79496c1f"). InnerVolumeSpecName "kube-api-access-6rmmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.681876 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-utilities\") pod \"00f163ab-337a-49bf-b3db-ca0c79496c1f\" (UID: \"00f163ab-337a-49bf-b3db-ca0c79496c1f\") " Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.683435 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-utilities" (OuterVolumeSpecName: "utilities") pod "00f163ab-337a-49bf-b3db-ca0c79496c1f" (UID: "00f163ab-337a-49bf-b3db-ca0c79496c1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.683553 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rmmh\" (UniqueName: \"kubernetes.io/projected/00f163ab-337a-49bf-b3db-ca0c79496c1f-kube-api-access-6rmmh\") on node \"crc\" DevicePath \"\"" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.691820 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00f163ab-337a-49bf-b3db-ca0c79496c1f" (UID: "00f163ab-337a-49bf-b3db-ca0c79496c1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.784849 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.784898 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f163ab-337a-49bf-b3db-ca0c79496c1f-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.926146 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-98gkb"] Dec 09 17:22:02 crc kubenswrapper[4954]: I1209 17:22:02.930980 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-98gkb"] Dec 09 17:22:04 crc kubenswrapper[4954]: I1209 17:22:04.128805 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00f163ab-337a-49bf-b3db-ca0c79496c1f" path="/var/lib/kubelet/pods/00f163ab-337a-49bf-b3db-ca0c79496c1f/volumes" Dec 09 17:22:10 crc kubenswrapper[4954]: I1209 17:22:10.921906 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6b9d7ddc4d-l8z78" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.000981 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x2qqj"] Dec 09 17:22:29 crc kubenswrapper[4954]: E1209 17:22:29.001860 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerName="extract-content" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.001876 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerName="extract-content" Dec 09 17:22:29 crc kubenswrapper[4954]: E1209 17:22:29.001900 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerName="registry-server" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.001908 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerName="registry-server" Dec 09 17:22:29 crc kubenswrapper[4954]: E1209 17:22:29.001921 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerName="extract-utilities" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.001931 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerName="extract-utilities" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.002076 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="00f163ab-337a-49bf-b3db-ca0c79496c1f" containerName="registry-server" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.006390 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.025901 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x2qqj"] Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.096180 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-catalog-content\") pod \"certified-operators-x2qqj\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.096251 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf2mm\" (UniqueName: \"kubernetes.io/projected/ea5afd95-8432-40b7-9e1a-b38360fdafdf-kube-api-access-sf2mm\") pod \"certified-operators-x2qqj\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.096328 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-utilities\") pod \"certified-operators-x2qqj\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.198150 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-catalog-content\") pod \"certified-operators-x2qqj\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.198207 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf2mm\" (UniqueName: \"kubernetes.io/projected/ea5afd95-8432-40b7-9e1a-b38360fdafdf-kube-api-access-sf2mm\") pod \"certified-operators-x2qqj\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.198255 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-utilities\") pod \"certified-operators-x2qqj\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.198740 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-catalog-content\") pod \"certified-operators-x2qqj\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.198751 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-utilities\") pod \"certified-operators-x2qqj\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.226784 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf2mm\" (UniqueName: \"kubernetes.io/projected/ea5afd95-8432-40b7-9e1a-b38360fdafdf-kube-api-access-sf2mm\") pod \"certified-operators-x2qqj\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.326975 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:29 crc kubenswrapper[4954]: I1209 17:22:29.781433 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x2qqj"] Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.230563 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7fb7cc7845-b4jd8" Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.783990 4954 generic.go:334] "Generic (PLEG): container finished" podID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerID="11b6b6b85ff0fca067fbffeec137360749daaddac1ff5570bf5f52d3e843ec86" exitCode=0 Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.784056 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2qqj" event={"ID":"ea5afd95-8432-40b7-9e1a-b38360fdafdf","Type":"ContainerDied","Data":"11b6b6b85ff0fca067fbffeec137360749daaddac1ff5570bf5f52d3e843ec86"} Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.784107 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2qqj" event={"ID":"ea5afd95-8432-40b7-9e1a-b38360fdafdf","Type":"ContainerStarted","Data":"7d3a707df0ff3199d95cb34939ec38e96cba9fe60642cc4f11a33576ed383ce2"} Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.974617 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-t56bj"] Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.977789 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.984106 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.984266 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-lfszs" Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.986372 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.989481 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg"] Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.991051 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" Dec 09 17:22:30 crc kubenswrapper[4954]: I1209 17:22:30.993382 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.006394 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg"] Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.078481 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-wl7gt"] Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.082348 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.085188 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.085423 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.085615 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-8ctrl" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.085974 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.096066 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-rtqnp"] Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.097447 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.102483 4954 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.115639 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-rtqnp"] Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.164548 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-metrics-certs\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.164621 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcn26\" (UniqueName: \"kubernetes.io/projected/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-kube-api-access-rcn26\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.164652 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-frr-conf\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.164678 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-frr-startup\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.164742 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-frr-sockets\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.164773 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-reloader\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.164819 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-metrics\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.164865 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khscx\" (UniqueName: \"kubernetes.io/projected/4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6-kube-api-access-khscx\") pod \"frr-k8s-webhook-server-7fcb986d4-f85kg\" (UID: \"4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.164916 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-f85kg\" (UID: \"4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.266729 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa6649f3-3933-4395-a345-f1a61a8616b1-cert\") pod \"controller-f8648f98b-rtqnp\" (UID: \"fa6649f3-3933-4395-a345-f1a61a8616b1\") " pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.266807 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-metrics\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.266855 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct5qg\" (UniqueName: \"kubernetes.io/projected/fa6649f3-3933-4395-a345-f1a61a8616b1-kube-api-access-ct5qg\") pod \"controller-f8648f98b-rtqnp\" (UID: \"fa6649f3-3933-4395-a345-f1a61a8616b1\") " pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.266882 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khscx\" (UniqueName: \"kubernetes.io/projected/4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6-kube-api-access-khscx\") pod \"frr-k8s-webhook-server-7fcb986d4-f85kg\" (UID: \"4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.266920 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/87be1fd5-5517-4e21-b5df-807a44761f4e-metallb-excludel2\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.266967 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-f85kg\" (UID: \"4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267003 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-metrics-certs\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267040 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-metrics-certs\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267061 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcn26\" (UniqueName: \"kubernetes.io/projected/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-kube-api-access-rcn26\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267091 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-memberlist\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267112 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2rph\" (UniqueName: \"kubernetes.io/projected/87be1fd5-5517-4e21-b5df-807a44761f4e-kube-api-access-x2rph\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267133 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-frr-conf\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267154 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-frr-startup\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267183 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa6649f3-3933-4395-a345-f1a61a8616b1-metrics-certs\") pod \"controller-f8648f98b-rtqnp\" (UID: \"fa6649f3-3933-4395-a345-f1a61a8616b1\") " pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267248 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-frr-sockets\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267280 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-reloader\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: E1209 17:22:31.267308 4954 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 09 17:22:31 crc kubenswrapper[4954]: E1209 17:22:31.267365 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-metrics-certs podName:b6f3c78e-58d6-4bba-a14e-85e3acd23e24 nodeName:}" failed. No retries permitted until 2025-12-09 17:22:31.767348133 +0000 UTC m=+1548.155521953 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-metrics-certs") pod "frr-k8s-t56bj" (UID: "b6f3c78e-58d6-4bba-a14e-85e3acd23e24") : secret "frr-k8s-certs-secret" not found Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267421 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-metrics\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267722 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-reloader\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.267786 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-frr-conf\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.268047 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-frr-sockets\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.268351 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-frr-startup\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.275042 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-f85kg\" (UID: \"4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.286927 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khscx\" (UniqueName: \"kubernetes.io/projected/4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6-kube-api-access-khscx\") pod \"frr-k8s-webhook-server-7fcb986d4-f85kg\" (UID: \"4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.294015 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcn26\" (UniqueName: \"kubernetes.io/projected/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-kube-api-access-rcn26\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.309103 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.368851 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-metrics-certs\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.368967 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-memberlist\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.370158 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2rph\" (UniqueName: \"kubernetes.io/projected/87be1fd5-5517-4e21-b5df-807a44761f4e-kube-api-access-x2rph\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.370239 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa6649f3-3933-4395-a345-f1a61a8616b1-metrics-certs\") pod \"controller-f8648f98b-rtqnp\" (UID: \"fa6649f3-3933-4395-a345-f1a61a8616b1\") " pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: E1209 17:22:31.369158 4954 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 09 17:22:31 crc kubenswrapper[4954]: E1209 17:22:31.370522 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-memberlist podName:87be1fd5-5517-4e21-b5df-807a44761f4e nodeName:}" failed. No retries permitted until 2025-12-09 17:22:31.870497843 +0000 UTC m=+1548.258671663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-memberlist") pod "speaker-wl7gt" (UID: "87be1fd5-5517-4e21-b5df-807a44761f4e") : secret "metallb-memberlist" not found Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.370793 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa6649f3-3933-4395-a345-f1a61a8616b1-cert\") pod \"controller-f8648f98b-rtqnp\" (UID: \"fa6649f3-3933-4395-a345-f1a61a8616b1\") " pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.371075 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct5qg\" (UniqueName: \"kubernetes.io/projected/fa6649f3-3933-4395-a345-f1a61a8616b1-kube-api-access-ct5qg\") pod \"controller-f8648f98b-rtqnp\" (UID: \"fa6649f3-3933-4395-a345-f1a61a8616b1\") " pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.371170 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/87be1fd5-5517-4e21-b5df-807a44761f4e-metallb-excludel2\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.372152 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/87be1fd5-5517-4e21-b5df-807a44761f4e-metallb-excludel2\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.373215 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-metrics-certs\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.377255 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa6649f3-3933-4395-a345-f1a61a8616b1-cert\") pod \"controller-f8648f98b-rtqnp\" (UID: \"fa6649f3-3933-4395-a345-f1a61a8616b1\") " pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.377501 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fa6649f3-3933-4395-a345-f1a61a8616b1-metrics-certs\") pod \"controller-f8648f98b-rtqnp\" (UID: \"fa6649f3-3933-4395-a345-f1a61a8616b1\") " pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.387383 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct5qg\" (UniqueName: \"kubernetes.io/projected/fa6649f3-3933-4395-a345-f1a61a8616b1-kube-api-access-ct5qg\") pod \"controller-f8648f98b-rtqnp\" (UID: \"fa6649f3-3933-4395-a345-f1a61a8616b1\") " pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.402579 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2rph\" (UniqueName: \"kubernetes.io/projected/87be1fd5-5517-4e21-b5df-807a44761f4e-kube-api-access-x2rph\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.419272 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.777155 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-metrics-certs\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.788095 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6f3c78e-58d6-4bba-a14e-85e3acd23e24-metrics-certs\") pod \"frr-k8s-t56bj\" (UID: \"b6f3c78e-58d6-4bba-a14e-85e3acd23e24\") " pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.796568 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2qqj" event={"ID":"ea5afd95-8432-40b7-9e1a-b38360fdafdf","Type":"ContainerStarted","Data":"37206fedb434f56083dc19bb6c9e9ea4d534edb3942ba556f65aecb89d5c0af8"} Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.813946 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg"] Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.879356 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-memberlist\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:31 crc kubenswrapper[4954]: E1209 17:22:31.879540 4954 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 09 17:22:31 crc kubenswrapper[4954]: E1209 17:22:31.879641 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-memberlist podName:87be1fd5-5517-4e21-b5df-807a44761f4e nodeName:}" failed. No retries permitted until 2025-12-09 17:22:32.879618467 +0000 UTC m=+1549.267792287 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-memberlist") pod "speaker-wl7gt" (UID: "87be1fd5-5517-4e21-b5df-807a44761f4e") : secret "metallb-memberlist" not found Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.899958 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:31 crc kubenswrapper[4954]: I1209 17:22:31.917771 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-rtqnp"] Dec 09 17:22:31 crc kubenswrapper[4954]: W1209 17:22:31.928978 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa6649f3_3933_4395_a345_f1a61a8616b1.slice/crio-b36e91b47b80b83e7712833d87a76a09433725bcadc5485db68288d90b9fcc93 WatchSource:0}: Error finding container b36e91b47b80b83e7712833d87a76a09433725bcadc5485db68288d90b9fcc93: Status 404 returned error can't find the container with id b36e91b47b80b83e7712833d87a76a09433725bcadc5485db68288d90b9fcc93 Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.805046 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t56bj" event={"ID":"b6f3c78e-58d6-4bba-a14e-85e3acd23e24","Type":"ContainerStarted","Data":"174f826b4178e4b7887f887baf3d77235036bf42753cb92d450147e511f7bab2"} Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.807071 4954 generic.go:334] "Generic (PLEG): container finished" podID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerID="37206fedb434f56083dc19bb6c9e9ea4d534edb3942ba556f65aecb89d5c0af8" exitCode=0 Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.807135 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2qqj" event={"ID":"ea5afd95-8432-40b7-9e1a-b38360fdafdf","Type":"ContainerDied","Data":"37206fedb434f56083dc19bb6c9e9ea4d534edb3942ba556f65aecb89d5c0af8"} Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.809745 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rtqnp" event={"ID":"fa6649f3-3933-4395-a345-f1a61a8616b1","Type":"ContainerStarted","Data":"20e4f43bae91d21ee0101e41618e008f7f5d311d525491601dcc8ab641bbc07c"} Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.809791 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rtqnp" event={"ID":"fa6649f3-3933-4395-a345-f1a61a8616b1","Type":"ContainerStarted","Data":"362945333ee41199ddaebac65ef163488277495153cba921367d1748b107ddf8"} Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.809806 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rtqnp" event={"ID":"fa6649f3-3933-4395-a345-f1a61a8616b1","Type":"ContainerStarted","Data":"b36e91b47b80b83e7712833d87a76a09433725bcadc5485db68288d90b9fcc93"} Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.810692 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.812990 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" event={"ID":"4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6","Type":"ContainerStarted","Data":"f241ea11e2237c33123958f7ed570748e2f754074ed9f85cfe1872bfc3af1f56"} Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.858785 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-rtqnp" podStartSLOduration=1.858764405 podStartE2EDuration="1.858764405s" podCreationTimestamp="2025-12-09 17:22:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:22:32.853385067 +0000 UTC m=+1549.241558897" watchObservedRunningTime="2025-12-09 17:22:32.858764405 +0000 UTC m=+1549.246938225" Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.899940 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-memberlist\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.907086 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/87be1fd5-5517-4e21-b5df-807a44761f4e-memberlist\") pod \"speaker-wl7gt\" (UID: \"87be1fd5-5517-4e21-b5df-807a44761f4e\") " pod="metallb-system/speaker-wl7gt" Dec 09 17:22:32 crc kubenswrapper[4954]: I1209 17:22:32.912457 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wl7gt" Dec 09 17:22:32 crc kubenswrapper[4954]: W1209 17:22:32.948499 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87be1fd5_5517_4e21_b5df_807a44761f4e.slice/crio-23b4c5b77faffde973519e8bc31ae3c3eb7baf01d5656de2dcd6e183ecaf9b1b WatchSource:0}: Error finding container 23b4c5b77faffde973519e8bc31ae3c3eb7baf01d5656de2dcd6e183ecaf9b1b: Status 404 returned error can't find the container with id 23b4c5b77faffde973519e8bc31ae3c3eb7baf01d5656de2dcd6e183ecaf9b1b Dec 09 17:22:33 crc kubenswrapper[4954]: I1209 17:22:33.838336 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2qqj" event={"ID":"ea5afd95-8432-40b7-9e1a-b38360fdafdf","Type":"ContainerStarted","Data":"0bf8c1bd0a1bf64959a3f1aec5e7f7beeb77ba2a812eec6a3a1033e5143bbb66"} Dec 09 17:22:33 crc kubenswrapper[4954]: I1209 17:22:33.846862 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wl7gt" event={"ID":"87be1fd5-5517-4e21-b5df-807a44761f4e","Type":"ContainerStarted","Data":"3983ea24853c9feee0823b3726e22292d3ffcfae7c53936bbac030629a2aa99d"} Dec 09 17:22:33 crc kubenswrapper[4954]: I1209 17:22:33.846920 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wl7gt" event={"ID":"87be1fd5-5517-4e21-b5df-807a44761f4e","Type":"ContainerStarted","Data":"be5bac13a6e9ed8c565e9c9ee87ed89b6652a5371f91304743bb59b3836f2db5"} Dec 09 17:22:33 crc kubenswrapper[4954]: I1209 17:22:33.846934 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wl7gt" event={"ID":"87be1fd5-5517-4e21-b5df-807a44761f4e","Type":"ContainerStarted","Data":"23b4c5b77faffde973519e8bc31ae3c3eb7baf01d5656de2dcd6e183ecaf9b1b"} Dec 09 17:22:33 crc kubenswrapper[4954]: I1209 17:22:33.847261 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-wl7gt" Dec 09 17:22:33 crc kubenswrapper[4954]: I1209 17:22:33.883574 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x2qqj" podStartSLOduration=3.380557988 podStartE2EDuration="5.883556419s" podCreationTimestamp="2025-12-09 17:22:28 +0000 UTC" firstStartedPulling="2025-12-09 17:22:30.786886203 +0000 UTC m=+1547.175060023" lastFinishedPulling="2025-12-09 17:22:33.289884634 +0000 UTC m=+1549.678058454" observedRunningTime="2025-12-09 17:22:33.878367317 +0000 UTC m=+1550.266541137" watchObservedRunningTime="2025-12-09 17:22:33.883556419 +0000 UTC m=+1550.271730239" Dec 09 17:22:33 crc kubenswrapper[4954]: I1209 17:22:33.913751 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-wl7gt" podStartSLOduration=2.913730971 podStartE2EDuration="2.913730971s" podCreationTimestamp="2025-12-09 17:22:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:22:33.907054713 +0000 UTC m=+1550.295228543" watchObservedRunningTime="2025-12-09 17:22:33.913730971 +0000 UTC m=+1550.301904791" Dec 09 17:22:39 crc kubenswrapper[4954]: I1209 17:22:39.327205 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:39 crc kubenswrapper[4954]: I1209 17:22:39.328889 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:39 crc kubenswrapper[4954]: I1209 17:22:39.387577 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:39 crc kubenswrapper[4954]: I1209 17:22:39.967421 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:40 crc kubenswrapper[4954]: I1209 17:22:40.028634 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x2qqj"] Dec 09 17:22:40 crc kubenswrapper[4954]: I1209 17:22:40.913049 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" event={"ID":"4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6","Type":"ContainerStarted","Data":"089903f455a3f60fe42d38bfa37cd43ea736e877e31b040660ce7cdff402c4e8"} Dec 09 17:22:40 crc kubenswrapper[4954]: I1209 17:22:40.913548 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" Dec 09 17:22:40 crc kubenswrapper[4954]: I1209 17:22:40.914967 4954 generic.go:334] "Generic (PLEG): container finished" podID="b6f3c78e-58d6-4bba-a14e-85e3acd23e24" containerID="eecfeb3f6c483a4cc42cb9903e77374e78cf2aed68e335e81cf28d1155f0f26e" exitCode=0 Dec 09 17:22:40 crc kubenswrapper[4954]: I1209 17:22:40.915011 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t56bj" event={"ID":"b6f3c78e-58d6-4bba-a14e-85e3acd23e24","Type":"ContainerDied","Data":"eecfeb3f6c483a4cc42cb9903e77374e78cf2aed68e335e81cf28d1155f0f26e"} Dec 09 17:22:40 crc kubenswrapper[4954]: I1209 17:22:40.955097 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" podStartSLOduration=3.036234836 podStartE2EDuration="10.955076227s" podCreationTimestamp="2025-12-09 17:22:30 +0000 UTC" firstStartedPulling="2025-12-09 17:22:31.819446459 +0000 UTC m=+1548.207620279" lastFinishedPulling="2025-12-09 17:22:39.73828785 +0000 UTC m=+1556.126461670" observedRunningTime="2025-12-09 17:22:40.945951713 +0000 UTC m=+1557.334125583" watchObservedRunningTime="2025-12-09 17:22:40.955076227 +0000 UTC m=+1557.343250047" Dec 09 17:22:41 crc kubenswrapper[4954]: I1209 17:22:41.924720 4954 generic.go:334] "Generic (PLEG): container finished" podID="b6f3c78e-58d6-4bba-a14e-85e3acd23e24" containerID="9449d5d20a18e2286caa49e6d6386927d5507a9090d05d5628c080833588c4d2" exitCode=0 Dec 09 17:22:41 crc kubenswrapper[4954]: I1209 17:22:41.924839 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t56bj" event={"ID":"b6f3c78e-58d6-4bba-a14e-85e3acd23e24","Type":"ContainerDied","Data":"9449d5d20a18e2286caa49e6d6386927d5507a9090d05d5628c080833588c4d2"} Dec 09 17:22:41 crc kubenswrapper[4954]: I1209 17:22:41.924968 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x2qqj" podUID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerName="registry-server" containerID="cri-o://0bf8c1bd0a1bf64959a3f1aec5e7f7beeb77ba2a812eec6a3a1033e5143bbb66" gracePeriod=2 Dec 09 17:22:42 crc kubenswrapper[4954]: I1209 17:22:42.934134 4954 generic.go:334] "Generic (PLEG): container finished" podID="b6f3c78e-58d6-4bba-a14e-85e3acd23e24" containerID="9cc5dfdaa0d5874718617d0ff0f29301cf8506c476eef9b8a69cd3497ba69644" exitCode=0 Dec 09 17:22:42 crc kubenswrapper[4954]: I1209 17:22:42.934221 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t56bj" event={"ID":"b6f3c78e-58d6-4bba-a14e-85e3acd23e24","Type":"ContainerDied","Data":"9cc5dfdaa0d5874718617d0ff0f29301cf8506c476eef9b8a69cd3497ba69644"} Dec 09 17:22:42 crc kubenswrapper[4954]: I1209 17:22:42.938340 4954 generic.go:334] "Generic (PLEG): container finished" podID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerID="0bf8c1bd0a1bf64959a3f1aec5e7f7beeb77ba2a812eec6a3a1033e5143bbb66" exitCode=0 Dec 09 17:22:42 crc kubenswrapper[4954]: I1209 17:22:42.938428 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2qqj" event={"ID":"ea5afd95-8432-40b7-9e1a-b38360fdafdf","Type":"ContainerDied","Data":"0bf8c1bd0a1bf64959a3f1aec5e7f7beeb77ba2a812eec6a3a1033e5143bbb66"} Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.504274 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.602646 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-catalog-content\") pod \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.602947 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf2mm\" (UniqueName: \"kubernetes.io/projected/ea5afd95-8432-40b7-9e1a-b38360fdafdf-kube-api-access-sf2mm\") pod \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.603052 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-utilities\") pod \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\" (UID: \"ea5afd95-8432-40b7-9e1a-b38360fdafdf\") " Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.603883 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-utilities" (OuterVolumeSpecName: "utilities") pod "ea5afd95-8432-40b7-9e1a-b38360fdafdf" (UID: "ea5afd95-8432-40b7-9e1a-b38360fdafdf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.608555 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea5afd95-8432-40b7-9e1a-b38360fdafdf-kube-api-access-sf2mm" (OuterVolumeSpecName: "kube-api-access-sf2mm") pod "ea5afd95-8432-40b7-9e1a-b38360fdafdf" (UID: "ea5afd95-8432-40b7-9e1a-b38360fdafdf"). InnerVolumeSpecName "kube-api-access-sf2mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.653280 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea5afd95-8432-40b7-9e1a-b38360fdafdf" (UID: "ea5afd95-8432-40b7-9e1a-b38360fdafdf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.706021 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.706061 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf2mm\" (UniqueName: \"kubernetes.io/projected/ea5afd95-8432-40b7-9e1a-b38360fdafdf-kube-api-access-sf2mm\") on node \"crc\" DevicePath \"\"" Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.706078 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea5afd95-8432-40b7-9e1a-b38360fdafdf-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.955121 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t56bj" event={"ID":"b6f3c78e-58d6-4bba-a14e-85e3acd23e24","Type":"ContainerStarted","Data":"ea8bffcabd0ad519ea8b45f6c82d7c46ba3e16506b58ed9f4972d17042f0da7f"} Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.955167 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t56bj" event={"ID":"b6f3c78e-58d6-4bba-a14e-85e3acd23e24","Type":"ContainerStarted","Data":"d459f6715fe549d270ddc42c6ba007038f59e15985bb4b5709aef891a09e686c"} Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.955177 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t56bj" event={"ID":"b6f3c78e-58d6-4bba-a14e-85e3acd23e24","Type":"ContainerStarted","Data":"3c3b22a2a50e3107dc4bc4f915789191a477261971832bf24a7dab4a0f6114b4"} Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.955189 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t56bj" event={"ID":"b6f3c78e-58d6-4bba-a14e-85e3acd23e24","Type":"ContainerStarted","Data":"187c11b08942fa9724d606ae8bb6f072c160d0e21cfc33dbf04f36010a77cd8e"} Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.955196 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t56bj" event={"ID":"b6f3c78e-58d6-4bba-a14e-85e3acd23e24","Type":"ContainerStarted","Data":"b2d4f7b68814f8672dc9f5a5cbfcfbcc65c7697589999f6a3afad28522b11b35"} Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.958452 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2qqj" event={"ID":"ea5afd95-8432-40b7-9e1a-b38360fdafdf","Type":"ContainerDied","Data":"7d3a707df0ff3199d95cb34939ec38e96cba9fe60642cc4f11a33576ed383ce2"} Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.958515 4954 scope.go:117] "RemoveContainer" containerID="0bf8c1bd0a1bf64959a3f1aec5e7f7beeb77ba2a812eec6a3a1033e5143bbb66" Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.958715 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2qqj" Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.990306 4954 scope.go:117] "RemoveContainer" containerID="37206fedb434f56083dc19bb6c9e9ea4d534edb3942ba556f65aecb89d5c0af8" Dec 09 17:22:43 crc kubenswrapper[4954]: I1209 17:22:43.996551 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x2qqj"] Dec 09 17:22:44 crc kubenswrapper[4954]: I1209 17:22:44.004852 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x2qqj"] Dec 09 17:22:44 crc kubenswrapper[4954]: I1209 17:22:44.009967 4954 scope.go:117] "RemoveContainer" containerID="11b6b6b85ff0fca067fbffeec137360749daaddac1ff5570bf5f52d3e843ec86" Dec 09 17:22:44 crc kubenswrapper[4954]: I1209 17:22:44.145026 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" path="/var/lib/kubelet/pods/ea5afd95-8432-40b7-9e1a-b38360fdafdf/volumes" Dec 09 17:22:44 crc kubenswrapper[4954]: I1209 17:22:44.970365 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t56bj" event={"ID":"b6f3c78e-58d6-4bba-a14e-85e3acd23e24","Type":"ContainerStarted","Data":"07f71aec886404cead912e6a30145b5fe76de79db578566abd92d96d061a4750"} Dec 09 17:22:44 crc kubenswrapper[4954]: I1209 17:22:44.970626 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:46 crc kubenswrapper[4954]: I1209 17:22:46.901243 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:46 crc kubenswrapper[4954]: I1209 17:22:46.947958 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-t56bj" Dec 09 17:22:46 crc kubenswrapper[4954]: I1209 17:22:46.974362 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-t56bj" podStartSLOduration=9.260602777999999 podStartE2EDuration="16.974344446s" podCreationTimestamp="2025-12-09 17:22:30 +0000 UTC" firstStartedPulling="2025-12-09 17:22:32.047794878 +0000 UTC m=+1548.435968698" lastFinishedPulling="2025-12-09 17:22:39.761536536 +0000 UTC m=+1556.149710366" observedRunningTime="2025-12-09 17:22:44.994762115 +0000 UTC m=+1561.382935955" watchObservedRunningTime="2025-12-09 17:22:46.974344446 +0000 UTC m=+1563.362518266" Dec 09 17:22:51 crc kubenswrapper[4954]: I1209 17:22:51.314923 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-f85kg" Dec 09 17:22:51 crc kubenswrapper[4954]: I1209 17:22:51.424161 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-rtqnp" Dec 09 17:22:52 crc kubenswrapper[4954]: I1209 17:22:52.917927 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-wl7gt" Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.801025 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-q6qbv"] Dec 09 17:22:55 crc kubenswrapper[4954]: E1209 17:22:55.801854 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerName="extract-content" Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.801874 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerName="extract-content" Dec 09 17:22:55 crc kubenswrapper[4954]: E1209 17:22:55.801894 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerName="extract-utilities" Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.801903 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerName="extract-utilities" Dec 09 17:22:55 crc kubenswrapper[4954]: E1209 17:22:55.801917 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerName="registry-server" Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.801926 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerName="registry-server" Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.802109 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea5afd95-8432-40b7-9e1a-b38360fdafdf" containerName="registry-server" Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.802832 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q6qbv" Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.805531 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-8dzkz" Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.805838 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.806044 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.857221 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-q6qbv"] Dec 09 17:22:55 crc kubenswrapper[4954]: I1209 17:22:55.955675 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxwpr\" (UniqueName: \"kubernetes.io/projected/7203e3a0-cf65-47cc-98ea-902b52d34363-kube-api-access-nxwpr\") pod \"openstack-operator-index-q6qbv\" (UID: \"7203e3a0-cf65-47cc-98ea-902b52d34363\") " pod="openstack-operators/openstack-operator-index-q6qbv" Dec 09 17:22:56 crc kubenswrapper[4954]: I1209 17:22:56.056796 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxwpr\" (UniqueName: \"kubernetes.io/projected/7203e3a0-cf65-47cc-98ea-902b52d34363-kube-api-access-nxwpr\") pod \"openstack-operator-index-q6qbv\" (UID: \"7203e3a0-cf65-47cc-98ea-902b52d34363\") " pod="openstack-operators/openstack-operator-index-q6qbv" Dec 09 17:22:56 crc kubenswrapper[4954]: I1209 17:22:56.080540 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxwpr\" (UniqueName: \"kubernetes.io/projected/7203e3a0-cf65-47cc-98ea-902b52d34363-kube-api-access-nxwpr\") pod \"openstack-operator-index-q6qbv\" (UID: \"7203e3a0-cf65-47cc-98ea-902b52d34363\") " pod="openstack-operators/openstack-operator-index-q6qbv" Dec 09 17:22:56 crc kubenswrapper[4954]: I1209 17:22:56.133560 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q6qbv" Dec 09 17:22:56 crc kubenswrapper[4954]: I1209 17:22:56.551843 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-q6qbv"] Dec 09 17:22:56 crc kubenswrapper[4954]: W1209 17:22:56.558764 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7203e3a0_cf65_47cc_98ea_902b52d34363.slice/crio-f2c1da5207dfc18d5c72ec000d4da09216badec74fea16788d20f9fc0180495d WatchSource:0}: Error finding container f2c1da5207dfc18d5c72ec000d4da09216badec74fea16788d20f9fc0180495d: Status 404 returned error can't find the container with id f2c1da5207dfc18d5c72ec000d4da09216badec74fea16788d20f9fc0180495d Dec 09 17:22:57 crc kubenswrapper[4954]: I1209 17:22:57.071949 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q6qbv" event={"ID":"7203e3a0-cf65-47cc-98ea-902b52d34363","Type":"ContainerStarted","Data":"f2c1da5207dfc18d5c72ec000d4da09216badec74fea16788d20f9fc0180495d"} Dec 09 17:22:57 crc kubenswrapper[4954]: I1209 17:22:57.968011 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-q6qbv"] Dec 09 17:22:58 crc kubenswrapper[4954]: I1209 17:22:58.387123 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-5t29f"] Dec 09 17:22:58 crc kubenswrapper[4954]: I1209 17:22:58.388789 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5t29f" Dec 09 17:22:58 crc kubenswrapper[4954]: I1209 17:22:58.409095 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5t29f"] Dec 09 17:22:58 crc kubenswrapper[4954]: I1209 17:22:58.503157 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j77m2\" (UniqueName: \"kubernetes.io/projected/428c0773-ebc5-4b31-b240-7aa8f6c2fcc9-kube-api-access-j77m2\") pod \"openstack-operator-index-5t29f\" (UID: \"428c0773-ebc5-4b31-b240-7aa8f6c2fcc9\") " pod="openstack-operators/openstack-operator-index-5t29f" Dec 09 17:22:58 crc kubenswrapper[4954]: I1209 17:22:58.605304 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j77m2\" (UniqueName: \"kubernetes.io/projected/428c0773-ebc5-4b31-b240-7aa8f6c2fcc9-kube-api-access-j77m2\") pod \"openstack-operator-index-5t29f\" (UID: \"428c0773-ebc5-4b31-b240-7aa8f6c2fcc9\") " pod="openstack-operators/openstack-operator-index-5t29f" Dec 09 17:22:58 crc kubenswrapper[4954]: I1209 17:22:58.641093 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j77m2\" (UniqueName: \"kubernetes.io/projected/428c0773-ebc5-4b31-b240-7aa8f6c2fcc9-kube-api-access-j77m2\") pod \"openstack-operator-index-5t29f\" (UID: \"428c0773-ebc5-4b31-b240-7aa8f6c2fcc9\") " pod="openstack-operators/openstack-operator-index-5t29f" Dec 09 17:22:58 crc kubenswrapper[4954]: I1209 17:22:58.723139 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5t29f" Dec 09 17:22:59 crc kubenswrapper[4954]: I1209 17:22:59.096477 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q6qbv" event={"ID":"7203e3a0-cf65-47cc-98ea-902b52d34363","Type":"ContainerStarted","Data":"7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9"} Dec 09 17:22:59 crc kubenswrapper[4954]: I1209 17:22:59.096653 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-q6qbv" podUID="7203e3a0-cf65-47cc-98ea-902b52d34363" containerName="registry-server" containerID="cri-o://7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9" gracePeriod=2 Dec 09 17:22:59 crc kubenswrapper[4954]: I1209 17:22:59.128117 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-q6qbv" podStartSLOduration=1.8339225369999999 podStartE2EDuration="4.128090089s" podCreationTimestamp="2025-12-09 17:22:55 +0000 UTC" firstStartedPulling="2025-12-09 17:22:56.562514944 +0000 UTC m=+1572.950688754" lastFinishedPulling="2025-12-09 17:22:58.856682486 +0000 UTC m=+1575.244856306" observedRunningTime="2025-12-09 17:22:59.112997977 +0000 UTC m=+1575.501171807" watchObservedRunningTime="2025-12-09 17:22:59.128090089 +0000 UTC m=+1575.516263909" Dec 09 17:22:59 crc kubenswrapper[4954]: I1209 17:22:59.224428 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5t29f"] Dec 09 17:22:59 crc kubenswrapper[4954]: W1209 17:22:59.227445 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod428c0773_ebc5_4b31_b240_7aa8f6c2fcc9.slice/crio-c4a2d60ba430e1e5ed6df193d705cc456f9e9f36617ec199de3b6938f03e6697 WatchSource:0}: Error finding container c4a2d60ba430e1e5ed6df193d705cc456f9e9f36617ec199de3b6938f03e6697: Status 404 returned error can't find the container with id c4a2d60ba430e1e5ed6df193d705cc456f9e9f36617ec199de3b6938f03e6697 Dec 09 17:22:59 crc kubenswrapper[4954]: I1209 17:22:59.540478 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q6qbv" Dec 09 17:22:59 crc kubenswrapper[4954]: I1209 17:22:59.621862 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxwpr\" (UniqueName: \"kubernetes.io/projected/7203e3a0-cf65-47cc-98ea-902b52d34363-kube-api-access-nxwpr\") pod \"7203e3a0-cf65-47cc-98ea-902b52d34363\" (UID: \"7203e3a0-cf65-47cc-98ea-902b52d34363\") " Dec 09 17:22:59 crc kubenswrapper[4954]: I1209 17:22:59.627722 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7203e3a0-cf65-47cc-98ea-902b52d34363-kube-api-access-nxwpr" (OuterVolumeSpecName: "kube-api-access-nxwpr") pod "7203e3a0-cf65-47cc-98ea-902b52d34363" (UID: "7203e3a0-cf65-47cc-98ea-902b52d34363"). InnerVolumeSpecName "kube-api-access-nxwpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:22:59 crc kubenswrapper[4954]: I1209 17:22:59.723557 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxwpr\" (UniqueName: \"kubernetes.io/projected/7203e3a0-cf65-47cc-98ea-902b52d34363-kube-api-access-nxwpr\") on node \"crc\" DevicePath \"\"" Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.106561 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5t29f" event={"ID":"428c0773-ebc5-4b31-b240-7aa8f6c2fcc9","Type":"ContainerStarted","Data":"6c0c7df1c18d68a97b2e14335d1b39850be44bc266bdac1e8705e0f0c8413be1"} Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.106642 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5t29f" event={"ID":"428c0773-ebc5-4b31-b240-7aa8f6c2fcc9","Type":"ContainerStarted","Data":"c4a2d60ba430e1e5ed6df193d705cc456f9e9f36617ec199de3b6938f03e6697"} Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.109110 4954 generic.go:334] "Generic (PLEG): container finished" podID="7203e3a0-cf65-47cc-98ea-902b52d34363" containerID="7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9" exitCode=0 Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.109173 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q6qbv" Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.109174 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q6qbv" event={"ID":"7203e3a0-cf65-47cc-98ea-902b52d34363","Type":"ContainerDied","Data":"7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9"} Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.109286 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q6qbv" event={"ID":"7203e3a0-cf65-47cc-98ea-902b52d34363","Type":"ContainerDied","Data":"f2c1da5207dfc18d5c72ec000d4da09216badec74fea16788d20f9fc0180495d"} Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.109318 4954 scope.go:117] "RemoveContainer" containerID="7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9" Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.127320 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-5t29f" podStartSLOduration=2.075744415 podStartE2EDuration="2.127302324s" podCreationTimestamp="2025-12-09 17:22:58 +0000 UTC" firstStartedPulling="2025-12-09 17:22:59.232563891 +0000 UTC m=+1575.620737711" lastFinishedPulling="2025-12-09 17:22:59.2841218 +0000 UTC m=+1575.672295620" observedRunningTime="2025-12-09 17:23:00.121677928 +0000 UTC m=+1576.509851748" watchObservedRunningTime="2025-12-09 17:23:00.127302324 +0000 UTC m=+1576.515476134" Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.160090 4954 scope.go:117] "RemoveContainer" containerID="7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9" Dec 09 17:23:00 crc kubenswrapper[4954]: E1209 17:23:00.160879 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9\": container with ID starting with 7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9 not found: ID does not exist" containerID="7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9" Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.160928 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9"} err="failed to get container status \"7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9\": rpc error: code = NotFound desc = could not find container \"7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9\": container with ID starting with 7f54118d60f2b55846cd61825832be90911ce27780f8545c92abda7e422667e9 not found: ID does not exist" Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.161672 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-q6qbv"] Dec 09 17:23:00 crc kubenswrapper[4954]: I1209 17:23:00.169813 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-q6qbv"] Dec 09 17:23:01 crc kubenswrapper[4954]: I1209 17:23:01.903264 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-t56bj" Dec 09 17:23:02 crc kubenswrapper[4954]: I1209 17:23:02.134914 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7203e3a0-cf65-47cc-98ea-902b52d34363" path="/var/lib/kubelet/pods/7203e3a0-cf65-47cc-98ea-902b52d34363/volumes" Dec 09 17:23:08 crc kubenswrapper[4954]: I1209 17:23:08.723828 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-5t29f" Dec 09 17:23:08 crc kubenswrapper[4954]: I1209 17:23:08.724797 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-5t29f" Dec 09 17:23:08 crc kubenswrapper[4954]: I1209 17:23:08.770274 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-5t29f" Dec 09 17:23:09 crc kubenswrapper[4954]: I1209 17:23:09.214757 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-5t29f" Dec 09 17:23:13 crc kubenswrapper[4954]: I1209 17:23:13.844148 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b"] Dec 09 17:23:13 crc kubenswrapper[4954]: E1209 17:23:13.845174 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7203e3a0-cf65-47cc-98ea-902b52d34363" containerName="registry-server" Dec 09 17:23:13 crc kubenswrapper[4954]: I1209 17:23:13.845197 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7203e3a0-cf65-47cc-98ea-902b52d34363" containerName="registry-server" Dec 09 17:23:13 crc kubenswrapper[4954]: I1209 17:23:13.845429 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="7203e3a0-cf65-47cc-98ea-902b52d34363" containerName="registry-server" Dec 09 17:23:13 crc kubenswrapper[4954]: I1209 17:23:13.846981 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:13 crc kubenswrapper[4954]: I1209 17:23:13.853946 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-xcc9z" Dec 09 17:23:13 crc kubenswrapper[4954]: I1209 17:23:13.856651 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b"] Dec 09 17:23:13 crc kubenswrapper[4954]: I1209 17:23:13.975415 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-bundle\") pod \"84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:13 crc kubenswrapper[4954]: I1209 17:23:13.976024 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzldb\" (UniqueName: \"kubernetes.io/projected/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-kube-api-access-pzldb\") pod \"84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:13 crc kubenswrapper[4954]: I1209 17:23:13.976359 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-util\") pod \"84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:14 crc kubenswrapper[4954]: I1209 17:23:14.078532 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-util\") pod \"84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:14 crc kubenswrapper[4954]: I1209 17:23:14.079049 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-bundle\") pod \"84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:14 crc kubenswrapper[4954]: I1209 17:23:14.079920 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzldb\" (UniqueName: \"kubernetes.io/projected/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-kube-api-access-pzldb\") pod \"84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:14 crc kubenswrapper[4954]: I1209 17:23:14.079831 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-bundle\") pod \"84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:14 crc kubenswrapper[4954]: I1209 17:23:14.079775 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-util\") pod \"84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:14 crc kubenswrapper[4954]: I1209 17:23:14.115282 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzldb\" (UniqueName: \"kubernetes.io/projected/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-kube-api-access-pzldb\") pod \"84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:14 crc kubenswrapper[4954]: I1209 17:23:14.184847 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:14 crc kubenswrapper[4954]: I1209 17:23:14.709416 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b"] Dec 09 17:23:14 crc kubenswrapper[4954]: W1209 17:23:14.719967 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd585c091_cf8b_41ef_8f4f_bfdfb004fd10.slice/crio-7f84dc4591461692af6c8e6d526c6aeab80057351fdc5544bd03d9e860f71d87 WatchSource:0}: Error finding container 7f84dc4591461692af6c8e6d526c6aeab80057351fdc5544bd03d9e860f71d87: Status 404 returned error can't find the container with id 7f84dc4591461692af6c8e6d526c6aeab80057351fdc5544bd03d9e860f71d87 Dec 09 17:23:15 crc kubenswrapper[4954]: I1209 17:23:15.257125 4954 generic.go:334] "Generic (PLEG): container finished" podID="d585c091-cf8b-41ef-8f4f-bfdfb004fd10" containerID="100e0ff92f932fe058a7c15e991c313375ffb4848a067e15777b3c12c4cc7dab" exitCode=0 Dec 09 17:23:15 crc kubenswrapper[4954]: I1209 17:23:15.257233 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" event={"ID":"d585c091-cf8b-41ef-8f4f-bfdfb004fd10","Type":"ContainerDied","Data":"100e0ff92f932fe058a7c15e991c313375ffb4848a067e15777b3c12c4cc7dab"} Dec 09 17:23:15 crc kubenswrapper[4954]: I1209 17:23:15.258334 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" event={"ID":"d585c091-cf8b-41ef-8f4f-bfdfb004fd10","Type":"ContainerStarted","Data":"7f84dc4591461692af6c8e6d526c6aeab80057351fdc5544bd03d9e860f71d87"} Dec 09 17:23:16 crc kubenswrapper[4954]: I1209 17:23:16.267963 4954 generic.go:334] "Generic (PLEG): container finished" podID="d585c091-cf8b-41ef-8f4f-bfdfb004fd10" containerID="42f0fd0ded470ec9a3a028be231e277de68c41b9d0bb1ba8f0cf4db6430d707d" exitCode=0 Dec 09 17:23:16 crc kubenswrapper[4954]: I1209 17:23:16.268227 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" event={"ID":"d585c091-cf8b-41ef-8f4f-bfdfb004fd10","Type":"ContainerDied","Data":"42f0fd0ded470ec9a3a028be231e277de68c41b9d0bb1ba8f0cf4db6430d707d"} Dec 09 17:23:17 crc kubenswrapper[4954]: I1209 17:23:17.286029 4954 generic.go:334] "Generic (PLEG): container finished" podID="d585c091-cf8b-41ef-8f4f-bfdfb004fd10" containerID="5b5203ce23ea41f4d3e0af11d0d63e2e735c2965d99b2b7b8d612126d64493c7" exitCode=0 Dec 09 17:23:17 crc kubenswrapper[4954]: I1209 17:23:17.286121 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" event={"ID":"d585c091-cf8b-41ef-8f4f-bfdfb004fd10","Type":"ContainerDied","Data":"5b5203ce23ea41f4d3e0af11d0d63e2e735c2965d99b2b7b8d612126d64493c7"} Dec 09 17:23:18 crc kubenswrapper[4954]: I1209 17:23:18.665062 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:18 crc kubenswrapper[4954]: I1209 17:23:18.768302 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-util\") pod \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " Dec 09 17:23:18 crc kubenswrapper[4954]: I1209 17:23:18.768470 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-bundle\") pod \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " Dec 09 17:23:18 crc kubenswrapper[4954]: I1209 17:23:18.768562 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzldb\" (UniqueName: \"kubernetes.io/projected/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-kube-api-access-pzldb\") pod \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\" (UID: \"d585c091-cf8b-41ef-8f4f-bfdfb004fd10\") " Dec 09 17:23:18 crc kubenswrapper[4954]: I1209 17:23:18.769488 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-bundle" (OuterVolumeSpecName: "bundle") pod "d585c091-cf8b-41ef-8f4f-bfdfb004fd10" (UID: "d585c091-cf8b-41ef-8f4f-bfdfb004fd10"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:23:18 crc kubenswrapper[4954]: I1209 17:23:18.775828 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-kube-api-access-pzldb" (OuterVolumeSpecName: "kube-api-access-pzldb") pod "d585c091-cf8b-41ef-8f4f-bfdfb004fd10" (UID: "d585c091-cf8b-41ef-8f4f-bfdfb004fd10"). InnerVolumeSpecName "kube-api-access-pzldb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:23:18 crc kubenswrapper[4954]: I1209 17:23:18.785076 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-util" (OuterVolumeSpecName: "util") pod "d585c091-cf8b-41ef-8f4f-bfdfb004fd10" (UID: "d585c091-cf8b-41ef-8f4f-bfdfb004fd10"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:23:18 crc kubenswrapper[4954]: I1209 17:23:18.871658 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzldb\" (UniqueName: \"kubernetes.io/projected/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-kube-api-access-pzldb\") on node \"crc\" DevicePath \"\"" Dec 09 17:23:18 crc kubenswrapper[4954]: I1209 17:23:18.871738 4954 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-util\") on node \"crc\" DevicePath \"\"" Dec 09 17:23:18 crc kubenswrapper[4954]: I1209 17:23:18.871753 4954 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d585c091-cf8b-41ef-8f4f-bfdfb004fd10-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:23:19 crc kubenswrapper[4954]: I1209 17:23:19.339814 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" event={"ID":"d585c091-cf8b-41ef-8f4f-bfdfb004fd10","Type":"ContainerDied","Data":"7f84dc4591461692af6c8e6d526c6aeab80057351fdc5544bd03d9e860f71d87"} Dec 09 17:23:19 crc kubenswrapper[4954]: I1209 17:23:19.339867 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f84dc4591461692af6c8e6d526c6aeab80057351fdc5544bd03d9e860f71d87" Dec 09 17:23:19 crc kubenswrapper[4954]: I1209 17:23:19.339909 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.084733 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98"] Dec 09 17:23:26 crc kubenswrapper[4954]: E1209 17:23:26.087224 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d585c091-cf8b-41ef-8f4f-bfdfb004fd10" containerName="pull" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.087324 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d585c091-cf8b-41ef-8f4f-bfdfb004fd10" containerName="pull" Dec 09 17:23:26 crc kubenswrapper[4954]: E1209 17:23:26.087396 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d585c091-cf8b-41ef-8f4f-bfdfb004fd10" containerName="extract" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.087466 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d585c091-cf8b-41ef-8f4f-bfdfb004fd10" containerName="extract" Dec 09 17:23:26 crc kubenswrapper[4954]: E1209 17:23:26.087536 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d585c091-cf8b-41ef-8f4f-bfdfb004fd10" containerName="util" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.087608 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d585c091-cf8b-41ef-8f4f-bfdfb004fd10" containerName="util" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.087893 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d585c091-cf8b-41ef-8f4f-bfdfb004fd10" containerName="extract" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.088881 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.091981 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-mbhk7" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.099760 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q4gk\" (UniqueName: \"kubernetes.io/projected/4a36b355-e0bf-4194-8c6a-fe39d7fa4b16-kube-api-access-9q4gk\") pod \"openstack-operator-controller-operator-5f5557f974-fsx98\" (UID: \"4a36b355-e0bf-4194-8c6a-fe39d7fa4b16\") " pod="openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.111681 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98"] Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.202291 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q4gk\" (UniqueName: \"kubernetes.io/projected/4a36b355-e0bf-4194-8c6a-fe39d7fa4b16-kube-api-access-9q4gk\") pod \"openstack-operator-controller-operator-5f5557f974-fsx98\" (UID: \"4a36b355-e0bf-4194-8c6a-fe39d7fa4b16\") " pod="openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.234295 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q4gk\" (UniqueName: \"kubernetes.io/projected/4a36b355-e0bf-4194-8c6a-fe39d7fa4b16-kube-api-access-9q4gk\") pod \"openstack-operator-controller-operator-5f5557f974-fsx98\" (UID: \"4a36b355-e0bf-4194-8c6a-fe39d7fa4b16\") " pod="openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.413957 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98" Dec 09 17:23:26 crc kubenswrapper[4954]: I1209 17:23:26.933911 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98"] Dec 09 17:23:27 crc kubenswrapper[4954]: I1209 17:23:27.406048 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98" event={"ID":"4a36b355-e0bf-4194-8c6a-fe39d7fa4b16","Type":"ContainerStarted","Data":"0755b35888d1aebd54e28a3c3ba20b75ec1221d2cac02c92d9c69ac844ad365f"} Dec 09 17:23:31 crc kubenswrapper[4954]: I1209 17:23:31.439288 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98" event={"ID":"4a36b355-e0bf-4194-8c6a-fe39d7fa4b16","Type":"ContainerStarted","Data":"97bb3a5c50647e15015481dd3e3fad7babed6194963c3daa0ea8189ca7a6effe"} Dec 09 17:23:31 crc kubenswrapper[4954]: I1209 17:23:31.439991 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98" Dec 09 17:23:31 crc kubenswrapper[4954]: I1209 17:23:31.471055 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98" podStartSLOduration=1.250027068 podStartE2EDuration="5.471035186s" podCreationTimestamp="2025-12-09 17:23:26 +0000 UTC" firstStartedPulling="2025-12-09 17:23:26.949408294 +0000 UTC m=+1603.337582114" lastFinishedPulling="2025-12-09 17:23:31.170416412 +0000 UTC m=+1607.558590232" observedRunningTime="2025-12-09 17:23:31.468561149 +0000 UTC m=+1607.856734989" watchObservedRunningTime="2025-12-09 17:23:31.471035186 +0000 UTC m=+1607.859209016" Dec 09 17:23:36 crc kubenswrapper[4954]: I1209 17:23:36.418712 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5f5557f974-fsx98" Dec 09 17:23:43 crc kubenswrapper[4954]: I1209 17:23:43.755982 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:23:43 crc kubenswrapper[4954]: I1209 17:23:43.756993 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.430454 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.433464 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.449817 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.452295 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.492211 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-m7bm6" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.504510 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-fn8c5" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.544631 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.566808 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.589935 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-8857b"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.595007 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.596485 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjwnv\" (UniqueName: \"kubernetes.io/projected/54373352-1448-443c-82eb-4bd28627acdf-kube-api-access-cjwnv\") pod \"cinder-operator-controller-manager-6c677c69b-6sfrq\" (UID: \"54373352-1448-443c-82eb-4bd28627acdf\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.596575 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44rpg\" (UniqueName: \"kubernetes.io/projected/e57eccbb-b484-45ad-a474-292da5bb6547-kube-api-access-44rpg\") pod \"barbican-operator-controller-manager-7d9dfd778-4pcjs\" (UID: \"e57eccbb-b484-45ad-a474-292da5bb6547\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.598049 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-t6l4n" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.631314 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.633049 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.635631 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-tcrx4" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.645923 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-8857b"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.663339 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.664958 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.667660 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.671353 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-rhf65" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.681617 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.684336 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.686876 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-hsd28" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.691692 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.698041 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkqkh\" (UniqueName: \"kubernetes.io/projected/41401dcd-a6d4-4e4b-93e9-bea5977d7e90-kube-api-access-hkqkh\") pod \"designate-operator-controller-manager-697fb699cf-8857b\" (UID: \"41401dcd-a6d4-4e4b-93e9-bea5977d7e90\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.698106 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44rpg\" (UniqueName: \"kubernetes.io/projected/e57eccbb-b484-45ad-a474-292da5bb6547-kube-api-access-44rpg\") pod \"barbican-operator-controller-manager-7d9dfd778-4pcjs\" (UID: \"e57eccbb-b484-45ad-a474-292da5bb6547\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.698200 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjwnv\" (UniqueName: \"kubernetes.io/projected/54373352-1448-443c-82eb-4bd28627acdf-kube-api-access-cjwnv\") pod \"cinder-operator-controller-manager-6c677c69b-6sfrq\" (UID: \"54373352-1448-443c-82eb-4bd28627acdf\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.700652 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.733690 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.735493 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.738436 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-n46ph" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.743974 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.746137 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.751156 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-sgdcb" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.751457 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.751805 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44rpg\" (UniqueName: \"kubernetes.io/projected/e57eccbb-b484-45ad-a474-292da5bb6547-kube-api-access-44rpg\") pod \"barbican-operator-controller-manager-7d9dfd778-4pcjs\" (UID: \"e57eccbb-b484-45ad-a474-292da5bb6547\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.760323 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjwnv\" (UniqueName: \"kubernetes.io/projected/54373352-1448-443c-82eb-4bd28627acdf-kube-api-access-cjwnv\") pod \"cinder-operator-controller-manager-6c677c69b-6sfrq\" (UID: \"54373352-1448-443c-82eb-4bd28627acdf\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.767577 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.807057 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrw8b\" (UniqueName: \"kubernetes.io/projected/3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8-kube-api-access-lrw8b\") pod \"glance-operator-controller-manager-5697bb5779-k5vlj\" (UID: \"3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.807103 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwfbr\" (UniqueName: \"kubernetes.io/projected/917a4c41-4cec-49ef-95f7-6cda64da9c44-kube-api-access-mwfbr\") pod \"heat-operator-controller-manager-5f64f6f8bb-8mpbt\" (UID: \"917a4c41-4cec-49ef-95f7-6cda64da9c44\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.807160 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.807254 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gxmx\" (UniqueName: \"kubernetes.io/projected/74b54335-7821-4cdc-a994-1dbe6c1653c9-kube-api-access-6gxmx\") pod \"horizon-operator-controller-manager-68c6d99b8f-28bcg\" (UID: \"74b54335-7821-4cdc-a994-1dbe6c1653c9\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.807316 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkqkh\" (UniqueName: \"kubernetes.io/projected/41401dcd-a6d4-4e4b-93e9-bea5977d7e90-kube-api-access-hkqkh\") pod \"designate-operator-controller-manager-697fb699cf-8857b\" (UID: \"41401dcd-a6d4-4e4b-93e9-bea5977d7e90\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.807341 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qmxw\" (UniqueName: \"kubernetes.io/projected/2b64512f-8bdc-42b5-9269-5ea7cd1eabf1-kube-api-access-7qmxw\") pod \"ironic-operator-controller-manager-967d97867-hlx7x\" (UID: \"2b64512f-8bdc-42b5-9269-5ea7cd1eabf1\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.807387 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c75f\" (UniqueName: \"kubernetes.io/projected/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-kube-api-access-6c75f\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.808043 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.828965 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.854218 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.857770 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkqkh\" (UniqueName: \"kubernetes.io/projected/41401dcd-a6d4-4e4b-93e9-bea5977d7e90-kube-api-access-hkqkh\") pod \"designate-operator-controller-manager-697fb699cf-8857b\" (UID: \"41401dcd-a6d4-4e4b-93e9-bea5977d7e90\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.866522 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.873506 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.878803 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.880804 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-ng25z" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.897363 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.899083 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.901865 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-xr9q2" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.911895 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gxmx\" (UniqueName: \"kubernetes.io/projected/74b54335-7821-4cdc-a994-1dbe6c1653c9-kube-api-access-6gxmx\") pod \"horizon-operator-controller-manager-68c6d99b8f-28bcg\" (UID: \"74b54335-7821-4cdc-a994-1dbe6c1653c9\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.911972 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qmxw\" (UniqueName: \"kubernetes.io/projected/2b64512f-8bdc-42b5-9269-5ea7cd1eabf1-kube-api-access-7qmxw\") pod \"ironic-operator-controller-manager-967d97867-hlx7x\" (UID: \"2b64512f-8bdc-42b5-9269-5ea7cd1eabf1\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.912014 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c75f\" (UniqueName: \"kubernetes.io/projected/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-kube-api-access-6c75f\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.912044 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwfbr\" (UniqueName: \"kubernetes.io/projected/917a4c41-4cec-49ef-95f7-6cda64da9c44-kube-api-access-mwfbr\") pod \"heat-operator-controller-manager-5f64f6f8bb-8mpbt\" (UID: \"917a4c41-4cec-49ef-95f7-6cda64da9c44\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.912108 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrw8b\" (UniqueName: \"kubernetes.io/projected/3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8-kube-api-access-lrw8b\") pod \"glance-operator-controller-manager-5697bb5779-k5vlj\" (UID: \"3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.912150 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:23:56 crc kubenswrapper[4954]: E1209 17:23:56.912381 4954 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 17:23:56 crc kubenswrapper[4954]: E1209 17:23:56.912479 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert podName:2ec773b3-dd1d-4662-9ebb-4be2993f76a2 nodeName:}" failed. No retries permitted until 2025-12-09 17:23:57.412444664 +0000 UTC m=+1633.800618484 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert") pod "infra-operator-controller-manager-78d48bff9d-cdpd9" (UID: "2ec773b3-dd1d-4662-9ebb-4be2993f76a2") : secret "infra-operator-webhook-server-cert" not found Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.913796 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.916920 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.926288 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-8j865" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.942175 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.950850 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c75f\" (UniqueName: \"kubernetes.io/projected/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-kube-api-access-6c75f\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.953355 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrw8b\" (UniqueName: \"kubernetes.io/projected/3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8-kube-api-access-lrw8b\") pod \"glance-operator-controller-manager-5697bb5779-k5vlj\" (UID: \"3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.953428 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb"] Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.954238 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.960026 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gxmx\" (UniqueName: \"kubernetes.io/projected/74b54335-7821-4cdc-a994-1dbe6c1653c9-kube-api-access-6gxmx\") pod \"horizon-operator-controller-manager-68c6d99b8f-28bcg\" (UID: \"74b54335-7821-4cdc-a994-1dbe6c1653c9\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.972894 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwfbr\" (UniqueName: \"kubernetes.io/projected/917a4c41-4cec-49ef-95f7-6cda64da9c44-kube-api-access-mwfbr\") pod \"heat-operator-controller-manager-5f64f6f8bb-8mpbt\" (UID: \"917a4c41-4cec-49ef-95f7-6cda64da9c44\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.982120 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.983672 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qmxw\" (UniqueName: \"kubernetes.io/projected/2b64512f-8bdc-42b5-9269-5ea7cd1eabf1-kube-api-access-7qmxw\") pod \"ironic-operator-controller-manager-967d97867-hlx7x\" (UID: \"2b64512f-8bdc-42b5-9269-5ea7cd1eabf1\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" Dec 09 17:23:56 crc kubenswrapper[4954]: I1209 17:23:56.989807 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.012114 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.016369 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.023096 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nzd9\" (UniqueName: \"kubernetes.io/projected/e84bf071-0cf4-4e92-9060-6cc7d30984db-kube-api-access-2nzd9\") pod \"keystone-operator-controller-manager-7765d96ddf-76tb6\" (UID: \"e84bf071-0cf4-4e92-9060-6cc7d30984db\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.023144 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.023882 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lt84\" (UniqueName: \"kubernetes.io/projected/6000ab00-d760-4e35-9ef6-47b878a8ef96-kube-api-access-7lt84\") pod \"manila-operator-controller-manager-5b5fd79c9c-g6ggb\" (UID: \"6000ab00-d760-4e35-9ef6-47b878a8ef96\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.032164 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-kswz4" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.114444 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.173786 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.176935 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jw5n\" (UniqueName: \"kubernetes.io/projected/8f82f473-cbee-4499-89b6-e6fbc32405aa-kube-api-access-8jw5n\") pod \"mariadb-operator-controller-manager-79c8c4686c-pfvmf\" (UID: \"8f82f473-cbee-4499-89b6-e6fbc32405aa\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.177003 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxxnn\" (UniqueName: \"kubernetes.io/projected/37d5d0b5-1e21-4b84-bd10-f4f738ef37a1-kube-api-access-xxxnn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-qsdpz\" (UID: \"37d5d0b5-1e21-4b84-bd10-f4f738ef37a1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.177040 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nzd9\" (UniqueName: \"kubernetes.io/projected/e84bf071-0cf4-4e92-9060-6cc7d30984db-kube-api-access-2nzd9\") pod \"keystone-operator-controller-manager-7765d96ddf-76tb6\" (UID: \"e84bf071-0cf4-4e92-9060-6cc7d30984db\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.177109 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lt84\" (UniqueName: \"kubernetes.io/projected/6000ab00-d760-4e35-9ef6-47b878a8ef96-kube-api-access-7lt84\") pod \"manila-operator-controller-manager-5b5fd79c9c-g6ggb\" (UID: \"6000ab00-d760-4e35-9ef6-47b878a8ef96\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.187203 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.245054 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.245644 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-9hrqh" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.278397 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jw5n\" (UniqueName: \"kubernetes.io/projected/8f82f473-cbee-4499-89b6-e6fbc32405aa-kube-api-access-8jw5n\") pod \"mariadb-operator-controller-manager-79c8c4686c-pfvmf\" (UID: \"8f82f473-cbee-4499-89b6-e6fbc32405aa\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.295484 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxxnn\" (UniqueName: \"kubernetes.io/projected/37d5d0b5-1e21-4b84-bd10-f4f738ef37a1-kube-api-access-xxxnn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-qsdpz\" (UID: \"37d5d0b5-1e21-4b84-bd10-f4f738ef37a1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.295580 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njs4z\" (UniqueName: \"kubernetes.io/projected/dc3d4dbd-0b2e-489e-b686-0195c372a917-kube-api-access-njs4z\") pod \"nova-operator-controller-manager-697bc559fc-mz7gh\" (UID: \"dc3d4dbd-0b2e-489e-b686-0195c372a917\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.323073 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.370415 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nzd9\" (UniqueName: \"kubernetes.io/projected/e84bf071-0cf4-4e92-9060-6cc7d30984db-kube-api-access-2nzd9\") pod \"keystone-operator-controller-manager-7765d96ddf-76tb6\" (UID: \"e84bf071-0cf4-4e92-9060-6cc7d30984db\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.390661 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.409397 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lt84\" (UniqueName: \"kubernetes.io/projected/6000ab00-d760-4e35-9ef6-47b878a8ef96-kube-api-access-7lt84\") pod \"manila-operator-controller-manager-5b5fd79c9c-g6ggb\" (UID: \"6000ab00-d760-4e35-9ef6-47b878a8ef96\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.412780 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxxnn\" (UniqueName: \"kubernetes.io/projected/37d5d0b5-1e21-4b84-bd10-f4f738ef37a1-kube-api-access-xxxnn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-qsdpz\" (UID: \"37d5d0b5-1e21-4b84-bd10-f4f738ef37a1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.421354 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.421761 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njs4z\" (UniqueName: \"kubernetes.io/projected/dc3d4dbd-0b2e-489e-b686-0195c372a917-kube-api-access-njs4z\") pod \"nova-operator-controller-manager-697bc559fc-mz7gh\" (UID: \"dc3d4dbd-0b2e-489e-b686-0195c372a917\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" Dec 09 17:23:57 crc kubenswrapper[4954]: E1209 17:23:57.422411 4954 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 17:23:57 crc kubenswrapper[4954]: E1209 17:23:57.422479 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert podName:2ec773b3-dd1d-4662-9ebb-4be2993f76a2 nodeName:}" failed. No retries permitted until 2025-12-09 17:23:58.422456061 +0000 UTC m=+1634.810629881 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert") pod "infra-operator-controller-manager-78d48bff9d-cdpd9" (UID: "2ec773b3-dd1d-4662-9ebb-4be2993f76a2") : secret "infra-operator-webhook-server-cert" not found Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.422941 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.432189 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jw5n\" (UniqueName: \"kubernetes.io/projected/8f82f473-cbee-4499-89b6-e6fbc32405aa-kube-api-access-8jw5n\") pod \"mariadb-operator-controller-manager-79c8c4686c-pfvmf\" (UID: \"8f82f473-cbee-4499-89b6-e6fbc32405aa\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.432732 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-bvlpl" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.537177 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.540105 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njs4z\" (UniqueName: \"kubernetes.io/projected/dc3d4dbd-0b2e-489e-b686-0195c372a917-kube-api-access-njs4z\") pod \"nova-operator-controller-manager-697bc559fc-mz7gh\" (UID: \"dc3d4dbd-0b2e-489e-b686-0195c372a917\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.540211 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.626112 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.626936 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.628725 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.642559 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tggrb\" (UniqueName: \"kubernetes.io/projected/2ba970e1-21a2-477e-947f-89de7bfbb6a6-kube-api-access-tggrb\") pod \"octavia-operator-controller-manager-998648c74-5jjz9\" (UID: \"2ba970e1-21a2-477e-947f-89de7bfbb6a6\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.651105 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.651913 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.686246 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.721419 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-jw8xq" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.722059 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.724943 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.725059 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.745283 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.748124 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.750492 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.752229 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tggrb\" (UniqueName: \"kubernetes.io/projected/2ba970e1-21a2-477e-947f-89de7bfbb6a6-kube-api-access-tggrb\") pod \"octavia-operator-controller-manager-998648c74-5jjz9\" (UID: \"2ba970e1-21a2-477e-947f-89de7bfbb6a6\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.752349 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7x6v\" (UniqueName: \"kubernetes.io/projected/281087c4-abfa-4537-80b5-130b03885954-kube-api-access-n7x6v\") pod \"ovn-operator-controller-manager-b6456fdb6-n7kmf\" (UID: \"281087c4-abfa-4537-80b5-130b03885954\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.753728 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-mnnr8" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.753996 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-bf9hh" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.755094 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.770719 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.815913 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.817472 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.823998 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-7d6vr" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.849196 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tggrb\" (UniqueName: \"kubernetes.io/projected/2ba970e1-21a2-477e-947f-89de7bfbb6a6-kube-api-access-tggrb\") pod \"octavia-operator-controller-manager-998648c74-5jjz9\" (UID: \"2ba970e1-21a2-477e-947f-89de7bfbb6a6\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.876872 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxpb6\" (UniqueName: \"kubernetes.io/projected/410c21d8-1970-411a-8704-db44a36c74cd-kube-api-access-mxpb6\") pod \"placement-operator-controller-manager-78f8948974-nvdgc\" (UID: \"410c21d8-1970-411a-8704-db44a36c74cd\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.876969 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxwdd\" (UniqueName: \"kubernetes.io/projected/3c0851fe-4209-4e7f-9622-dba385e6bc78-kube-api-access-gxwdd\") pod \"swift-operator-controller-manager-9d58d64bc-8ltxl\" (UID: \"3c0851fe-4209-4e7f-9622-dba385e6bc78\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.877163 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7x6v\" (UniqueName: \"kubernetes.io/projected/281087c4-abfa-4537-80b5-130b03885954-kube-api-access-n7x6v\") pod \"ovn-operator-controller-manager-b6456fdb6-n7kmf\" (UID: \"281087c4-abfa-4537-80b5-130b03885954\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.877466 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n4cq\" (UniqueName: \"kubernetes.io/projected/074bde7a-88ab-45ac-8a29-09f49834e4c8-kube-api-access-8n4cq\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.877502 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.898679 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.903777 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.920344 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.927762 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-wgm4f" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.935256 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7x6v\" (UniqueName: \"kubernetes.io/projected/281087c4-abfa-4537-80b5-130b03885954-kube-api-access-n7x6v\") pod \"ovn-operator-controller-manager-b6456fdb6-n7kmf\" (UID: \"281087c4-abfa-4537-80b5-130b03885954\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.963678 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.973667 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.975754 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-96qpl"] Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.979065 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.980666 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n4cq\" (UniqueName: \"kubernetes.io/projected/074bde7a-88ab-45ac-8a29-09f49834e4c8-kube-api-access-8n4cq\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.980712 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.980756 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxpb6\" (UniqueName: \"kubernetes.io/projected/410c21d8-1970-411a-8704-db44a36c74cd-kube-api-access-mxpb6\") pod \"placement-operator-controller-manager-78f8948974-nvdgc\" (UID: \"410c21d8-1970-411a-8704-db44a36c74cd\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.981890 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxwdd\" (UniqueName: \"kubernetes.io/projected/3c0851fe-4209-4e7f-9622-dba385e6bc78-kube-api-access-gxwdd\") pod \"swift-operator-controller-manager-9d58d64bc-8ltxl\" (UID: \"3c0851fe-4209-4e7f-9622-dba385e6bc78\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.981986 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9ht6\" (UniqueName: \"kubernetes.io/projected/93405692-e3d5-4a81-97db-170d70b6cbc3-kube-api-access-m9ht6\") pod \"telemetry-operator-controller-manager-796785f986-j59bx\" (UID: \"93405692-e3d5-4a81-97db-170d70b6cbc3\") " pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" Dec 09 17:23:57 crc kubenswrapper[4954]: E1209 17:23:57.983665 4954 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 17:23:57 crc kubenswrapper[4954]: E1209 17:23:57.983788 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert podName:074bde7a-88ab-45ac-8a29-09f49834e4c8 nodeName:}" failed. No retries permitted until 2025-12-09 17:23:58.483756632 +0000 UTC m=+1634.871930442 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fzcn6j" (UID: "074bde7a-88ab-45ac-8a29-09f49834e4c8") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.986911 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-x867c" Dec 09 17:23:57 crc kubenswrapper[4954]: I1209 17:23:57.987642 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-96qpl"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.001369 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.003764 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.012979 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.021046 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.023423 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxpb6\" (UniqueName: \"kubernetes.io/projected/410c21d8-1970-411a-8704-db44a36c74cd-kube-api-access-mxpb6\") pod \"placement-operator-controller-manager-78f8948974-nvdgc\" (UID: \"410c21d8-1970-411a-8704-db44a36c74cd\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.025166 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.027158 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-jbmzk" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.033321 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.038853 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n4cq\" (UniqueName: \"kubernetes.io/projected/074bde7a-88ab-45ac-8a29-09f49834e4c8-kube-api-access-8n4cq\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.038923 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.039319 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zhrwb" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.039362 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.042755 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxwdd\" (UniqueName: \"kubernetes.io/projected/3c0851fe-4209-4e7f-9622-dba385e6bc78-kube-api-access-gxwdd\") pod \"swift-operator-controller-manager-9d58d64bc-8ltxl\" (UID: \"3c0851fe-4209-4e7f-9622-dba385e6bc78\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.063832 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.069481 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.093984 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.104187 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz78s\" (UniqueName: \"kubernetes.io/projected/ce9db630-dee1-4c9f-b90e-31d263b8a213-kube-api-access-dz78s\") pod \"test-operator-controller-manager-5854674fcc-96qpl\" (UID: \"ce9db630-dee1-4c9f-b90e-31d263b8a213\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.107508 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.109006 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.109230 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4x47\" (UniqueName: \"kubernetes.io/projected/1f87c2e5-6b6b-4f67-b2f6-caf321fbe844-kube-api-access-p4x47\") pod \"watcher-operator-controller-manager-667bd8d554-4rrqn\" (UID: \"1f87c2e5-6b6b-4f67-b2f6-caf321fbe844\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.111616 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qwnd\" (UniqueName: \"kubernetes.io/projected/4ec17be0-419e-48e3-a00f-58741723b85d-kube-api-access-6qwnd\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.111771 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9ht6\" (UniqueName: \"kubernetes.io/projected/93405692-e3d5-4a81-97db-170d70b6cbc3-kube-api-access-m9ht6\") pod \"telemetry-operator-controller-manager-796785f986-j59bx\" (UID: \"93405692-e3d5-4a81-97db-170d70b6cbc3\") " pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.139471 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9ht6\" (UniqueName: \"kubernetes.io/projected/93405692-e3d5-4a81-97db-170d70b6cbc3-kube-api-access-m9ht6\") pod \"telemetry-operator-controller-manager-796785f986-j59bx\" (UID: \"93405692-e3d5-4a81-97db-170d70b6cbc3\") " pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.145231 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.215212 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.215278 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.215357 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4x47\" (UniqueName: \"kubernetes.io/projected/1f87c2e5-6b6b-4f67-b2f6-caf321fbe844-kube-api-access-p4x47\") pod \"watcher-operator-controller-manager-667bd8d554-4rrqn\" (UID: \"1f87c2e5-6b6b-4f67-b2f6-caf321fbe844\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.215460 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qwnd\" (UniqueName: \"kubernetes.io/projected/4ec17be0-419e-48e3-a00f-58741723b85d-kube-api-access-6qwnd\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.215564 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz78s\" (UniqueName: \"kubernetes.io/projected/ce9db630-dee1-4c9f-b90e-31d263b8a213-kube-api-access-dz78s\") pod \"test-operator-controller-manager-5854674fcc-96qpl\" (UID: \"ce9db630-dee1-4c9f-b90e-31d263b8a213\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.217001 4954 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.217067 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:23:58.717043155 +0000 UTC m=+1635.105216975 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "webhook-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.217343 4954 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.217392 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:23:58.717382286 +0000 UTC m=+1635.105556106 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "metrics-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.250999 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz78s\" (UniqueName: \"kubernetes.io/projected/ce9db630-dee1-4c9f-b90e-31d263b8a213-kube-api-access-dz78s\") pod \"test-operator-controller-manager-5854674fcc-96qpl\" (UID: \"ce9db630-dee1-4c9f-b90e-31d263b8a213\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.252734 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.254614 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.254666 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.254786 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.262550 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qwnd\" (UniqueName: \"kubernetes.io/projected/4ec17be0-419e-48e3-a00f-58741723b85d-kube-api-access-6qwnd\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.262550 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-6ttlv" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.275903 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4x47\" (UniqueName: \"kubernetes.io/projected/1f87c2e5-6b6b-4f67-b2f6-caf321fbe844-kube-api-access-p4x47\") pod \"watcher-operator-controller-manager-667bd8d554-4rrqn\" (UID: \"1f87c2e5-6b6b-4f67-b2f6-caf321fbe844\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.308890 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.311924 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.361839 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.423070 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.423635 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnl5f\" (UniqueName: \"kubernetes.io/projected/1fc4c8d7-5079-45d0-8a68-d2523e7b981f-kube-api-access-wnl5f\") pod \"rabbitmq-cluster-operator-manager-668c99d594-rcwxm\" (UID: \"1fc4c8d7-5079-45d0-8a68-d2523e7b981f\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.423744 4954 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.423791 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert podName:2ec773b3-dd1d-4662-9ebb-4be2993f76a2 nodeName:}" failed. No retries permitted until 2025-12-09 17:24:00.423774287 +0000 UTC m=+1636.811948107 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert") pod "infra-operator-controller-manager-78d48bff9d-cdpd9" (UID: "2ec773b3-dd1d-4662-9ebb-4be2993f76a2") : secret "infra-operator-webhook-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.526349 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnl5f\" (UniqueName: \"kubernetes.io/projected/1fc4c8d7-5079-45d0-8a68-d2523e7b981f-kube-api-access-wnl5f\") pod \"rabbitmq-cluster-operator-manager-668c99d594-rcwxm\" (UID: \"1fc4c8d7-5079-45d0-8a68-d2523e7b981f\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.526443 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.526613 4954 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.526663 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert podName:074bde7a-88ab-45ac-8a29-09f49834e4c8 nodeName:}" failed. No retries permitted until 2025-12-09 17:23:59.526645057 +0000 UTC m=+1635.914818877 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fzcn6j" (UID: "074bde7a-88ab-45ac-8a29-09f49834e4c8") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.546121 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnl5f\" (UniqueName: \"kubernetes.io/projected/1fc4c8d7-5079-45d0-8a68-d2523e7b981f-kube-api-access-wnl5f\") pod \"rabbitmq-cluster-operator-manager-668c99d594-rcwxm\" (UID: \"1fc4c8d7-5079-45d0-8a68-d2523e7b981f\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.708250 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.729270 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.729324 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.729570 4954 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.729712 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:23:59.729678854 +0000 UTC m=+1636.117852844 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "webhook-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.729734 4954 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: E1209 17:23:58.729830 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:23:59.729805518 +0000 UTC m=+1636.117979518 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "metrics-server-cert" not found Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.864536 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" event={"ID":"54373352-1448-443c-82eb-4bd28627acdf","Type":"ContainerStarted","Data":"4a63b487c28b1f941fce90036035d235a42312910367b8a7e581c17d334fb659"} Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.874872 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" event={"ID":"e57eccbb-b484-45ad-a474-292da5bb6547","Type":"ContainerStarted","Data":"4caca8ce5c08791aeaefada524621ea6f488f50111b0a45db42f46cfc0fcca77"} Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.896318 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-8857b"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.939024 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj"] Dec 09 17:23:58 crc kubenswrapper[4954]: I1209 17:23:58.954802 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt"] Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.194829 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg"] Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.214976 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb"] Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.237324 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x"] Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.280681 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf"] Dec 09 17:23:59 crc kubenswrapper[4954]: W1209 17:23:59.407222 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod281087c4_abfa_4537_80b5_130b03885954.slice/crio-8ec8af086111afaa8903df082d80f83a0cef94fb8afd4e9b6e48cd1b513501f7 WatchSource:0}: Error finding container 8ec8af086111afaa8903df082d80f83a0cef94fb8afd4e9b6e48cd1b513501f7: Status 404 returned error can't find the container with id 8ec8af086111afaa8903df082d80f83a0cef94fb8afd4e9b6e48cd1b513501f7 Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.448487 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf"] Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.458727 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz"] Dec 09 17:23:59 crc kubenswrapper[4954]: W1209 17:23:59.475014 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f82f473_cbee_4499_89b6_e6fbc32405aa.slice/crio-c4f919de58ce0b2b1dd53f595238ac6284e9036a1d2c5b68e767c1d57c1c3581 WatchSource:0}: Error finding container c4f919de58ce0b2b1dd53f595238ac6284e9036a1d2c5b68e767c1d57c1c3581: Status 404 returned error can't find the container with id c4f919de58ce0b2b1dd53f595238ac6284e9036a1d2c5b68e767c1d57c1c3581 Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.495481 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx"] Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.505453 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh"] Dec 09 17:23:59 crc kubenswrapper[4954]: W1209 17:23:59.517241 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc3d4dbd_0b2e_489e_b686_0195c372a917.slice/crio-c0e871177426faf5f8b06988d934e315d671dc97f12cf7712247023f9b262d23 WatchSource:0}: Error finding container c0e871177426faf5f8b06988d934e315d671dc97f12cf7712247023f9b262d23: Status 404 returned error can't find the container with id c0e871177426faf5f8b06988d934e315d671dc97f12cf7712247023f9b262d23 Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.552308 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.552704 4954 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.552786 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert podName:074bde7a-88ab-45ac-8a29-09f49834e4c8 nodeName:}" failed. No retries permitted until 2025-12-09 17:24:01.552760011 +0000 UTC m=+1637.940933831 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fzcn6j" (UID: "074bde7a-88ab-45ac-8a29-09f49834e4c8") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.646120 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-96qpl"] Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.656505 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6"] Dec 09 17:23:59 crc kubenswrapper[4954]: W1209 17:23:59.658380 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c0851fe_4209_4e7f_9622_dba385e6bc78.slice/crio-7703fd1e06fc4ae88a473f3ef99913dc49ee2bf9b98cd731ab5afba13afdf9cd WatchSource:0}: Error finding container 7703fd1e06fc4ae88a473f3ef99913dc49ee2bf9b98cd731ab5afba13afdf9cd: Status 404 returned error can't find the container with id 7703fd1e06fc4ae88a473f3ef99913dc49ee2bf9b98cd731ab5afba13afdf9cd Dec 09 17:23:59 crc kubenswrapper[4954]: W1209 17:23:59.668375 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode84bf071_0cf4_4e92_9060_6cc7d30984db.slice/crio-714e9aef4a7324c5ec364eef59e1c1d83b3b1954aa70f82c3e164430d239c513 WatchSource:0}: Error finding container 714e9aef4a7324c5ec364eef59e1c1d83b3b1954aa70f82c3e164430d239c513: Status 404 returned error can't find the container with id 714e9aef4a7324c5ec364eef59e1c1d83b3b1954aa70f82c3e164430d239c513 Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.670502 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl"] Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.682229 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc"] Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.690356 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn"] Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.697427 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2nzd9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-76tb6_openstack-operators(e84bf071-0cf4-4e92-9060-6cc7d30984db): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.699528 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9"] Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.701069 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2nzd9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-76tb6_openstack-operators(e84bf071-0cf4-4e92-9060-6cc7d30984db): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.701237 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dz78s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-96qpl_openstack-operators(ce9db630-dee1-4c9f-b90e-31d263b8a213): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.702697 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" podUID="e84bf071-0cf4-4e92-9060-6cc7d30984db" Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.703653 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dz78s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-96qpl_openstack-operators(ce9db630-dee1-4c9f-b90e-31d263b8a213): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.703764 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tggrb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-5jjz9_openstack-operators(2ba970e1-21a2-477e-947f-89de7bfbb6a6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.704895 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" podUID="ce9db630-dee1-4c9f-b90e-31d263b8a213" Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.705995 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tggrb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-5jjz9_openstack-operators(2ba970e1-21a2-477e-947f-89de7bfbb6a6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.707350 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" podUID="2ba970e1-21a2-477e-947f-89de7bfbb6a6" Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.735546 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm"] Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.752754 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wnl5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-rcwxm_openstack-operators(1fc4c8d7-5079-45d0-8a68-d2523e7b981f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.754262 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" podUID="1fc4c8d7-5079-45d0-8a68-d2523e7b981f" Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.755972 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.756058 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.756297 4954 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.756403 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:24:01.756376505 +0000 UTC m=+1638.144550325 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "webhook-server-cert" not found Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.756569 4954 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.756741 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:24:01.756714376 +0000 UTC m=+1638.144888196 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "metrics-server-cert" not found Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.934447 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" event={"ID":"917a4c41-4cec-49ef-95f7-6cda64da9c44","Type":"ContainerStarted","Data":"7938736a2c626d8c365e09c174aad419543d3391cfc9f1cde4f632baa422a46a"} Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.942980 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" event={"ID":"ce9db630-dee1-4c9f-b90e-31d263b8a213","Type":"ContainerStarted","Data":"b77fa08dd14f140d14389046b2cc08b4e2e6d18787a436c6b0a297851f6d2783"} Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.947998 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" event={"ID":"41401dcd-a6d4-4e4b-93e9-bea5977d7e90","Type":"ContainerStarted","Data":"efcf5973f57ad305ec4fd72a86c1aab3c0d53c3c7ad4da9c5bc22f85021a7a96"} Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.956007 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" podUID="ce9db630-dee1-4c9f-b90e-31d263b8a213" Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.958870 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" event={"ID":"dc3d4dbd-0b2e-489e-b686-0195c372a917","Type":"ContainerStarted","Data":"c0e871177426faf5f8b06988d934e315d671dc97f12cf7712247023f9b262d23"} Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.966050 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" event={"ID":"e84bf071-0cf4-4e92-9060-6cc7d30984db","Type":"ContainerStarted","Data":"714e9aef4a7324c5ec364eef59e1c1d83b3b1954aa70f82c3e164430d239c513"} Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.970560 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" event={"ID":"37d5d0b5-1e21-4b84-bd10-f4f738ef37a1","Type":"ContainerStarted","Data":"ec649a9962d5e183e23b9b4066cb7e0dbbf2271316cf33b45d76b85cc043bde4"} Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.976361 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" event={"ID":"3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8","Type":"ContainerStarted","Data":"cbf77e5edf9f8b4697ad2bda765c63c4e3fec49a3de91ac95b50c8e5c862698f"} Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.978894 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" event={"ID":"3c0851fe-4209-4e7f-9622-dba385e6bc78","Type":"ContainerStarted","Data":"7703fd1e06fc4ae88a473f3ef99913dc49ee2bf9b98cd731ab5afba13afdf9cd"} Dec 09 17:23:59 crc kubenswrapper[4954]: E1209 17:23:59.980288 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" podUID="e84bf071-0cf4-4e92-9060-6cc7d30984db" Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.995211 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" event={"ID":"281087c4-abfa-4537-80b5-130b03885954","Type":"ContainerStarted","Data":"8ec8af086111afaa8903df082d80f83a0cef94fb8afd4e9b6e48cd1b513501f7"} Dec 09 17:23:59 crc kubenswrapper[4954]: I1209 17:23:59.999414 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" event={"ID":"2ba970e1-21a2-477e-947f-89de7bfbb6a6","Type":"ContainerStarted","Data":"ce55c443da710fa9dd32aa2ea346a0cc7c6eb3786223b642bb83b891c48246aa"} Dec 09 17:24:00 crc kubenswrapper[4954]: E1209 17:24:00.002739 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" podUID="2ba970e1-21a2-477e-947f-89de7bfbb6a6" Dec 09 17:24:00 crc kubenswrapper[4954]: I1209 17:24:00.002887 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" event={"ID":"410c21d8-1970-411a-8704-db44a36c74cd","Type":"ContainerStarted","Data":"91ee754659b55aa3e838348df458d932ad61d802a73a77beb997e868fb105caa"} Dec 09 17:24:00 crc kubenswrapper[4954]: I1209 17:24:00.008927 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" event={"ID":"1f87c2e5-6b6b-4f67-b2f6-caf321fbe844","Type":"ContainerStarted","Data":"060a38282be0703b744377fb2484fb59eb6bca1f2c29f7495982d0b564d75594"} Dec 09 17:24:00 crc kubenswrapper[4954]: I1209 17:24:00.012874 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" event={"ID":"1fc4c8d7-5079-45d0-8a68-d2523e7b981f","Type":"ContainerStarted","Data":"69a90894d0436a07332737d708af3f59aa8ec341f8f8291f2de1c598ec9d5da2"} Dec 09 17:24:00 crc kubenswrapper[4954]: I1209 17:24:00.014994 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" event={"ID":"6000ab00-d760-4e35-9ef6-47b878a8ef96","Type":"ContainerStarted","Data":"319f63be4d9b5e2f211c83fb7d6c30ddd4bc4207683350ef865b1e8469bfb5b9"} Dec 09 17:24:00 crc kubenswrapper[4954]: E1209 17:24:00.017000 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" podUID="1fc4c8d7-5079-45d0-8a68-d2523e7b981f" Dec 09 17:24:00 crc kubenswrapper[4954]: I1209 17:24:00.034329 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" event={"ID":"93405692-e3d5-4a81-97db-170d70b6cbc3","Type":"ContainerStarted","Data":"b171b3ef37cfa5d131b5b9b5693a044103d9fef1f5022911cfad5e8a1aac70bb"} Dec 09 17:24:00 crc kubenswrapper[4954]: I1209 17:24:00.038967 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" event={"ID":"2b64512f-8bdc-42b5-9269-5ea7cd1eabf1","Type":"ContainerStarted","Data":"1c9eefe510e04cd99b9742c6b809c4cd4bcb160e4cdbf3e4d2a81ec9cb2f1c44"} Dec 09 17:24:00 crc kubenswrapper[4954]: I1209 17:24:00.041333 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" event={"ID":"8f82f473-cbee-4499-89b6-e6fbc32405aa","Type":"ContainerStarted","Data":"c4f919de58ce0b2b1dd53f595238ac6284e9036a1d2c5b68e767c1d57c1c3581"} Dec 09 17:24:00 crc kubenswrapper[4954]: I1209 17:24:00.043585 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" event={"ID":"74b54335-7821-4cdc-a994-1dbe6c1653c9","Type":"ContainerStarted","Data":"4245b15a3c25fb78c41011b538524cb11eb83ce85366478662107ec9bdb99fca"} Dec 09 17:24:00 crc kubenswrapper[4954]: I1209 17:24:00.482377 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:24:00 crc kubenswrapper[4954]: E1209 17:24:00.482648 4954 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 17:24:00 crc kubenswrapper[4954]: E1209 17:24:00.482786 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert podName:2ec773b3-dd1d-4662-9ebb-4be2993f76a2 nodeName:}" failed. No retries permitted until 2025-12-09 17:24:04.482756354 +0000 UTC m=+1640.870930344 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert") pod "infra-operator-controller-manager-78d48bff9d-cdpd9" (UID: "2ec773b3-dd1d-4662-9ebb-4be2993f76a2") : secret "infra-operator-webhook-server-cert" not found Dec 09 17:24:01 crc kubenswrapper[4954]: E1209 17:24:01.058342 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" podUID="1fc4c8d7-5079-45d0-8a68-d2523e7b981f" Dec 09 17:24:01 crc kubenswrapper[4954]: E1209 17:24:01.059430 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" podUID="2ba970e1-21a2-477e-947f-89de7bfbb6a6" Dec 09 17:24:01 crc kubenswrapper[4954]: E1209 17:24:01.068582 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" podUID="e84bf071-0cf4-4e92-9060-6cc7d30984db" Dec 09 17:24:01 crc kubenswrapper[4954]: E1209 17:24:01.074581 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" podUID="ce9db630-dee1-4c9f-b90e-31d263b8a213" Dec 09 17:24:01 crc kubenswrapper[4954]: I1209 17:24:01.607808 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:24:01 crc kubenswrapper[4954]: E1209 17:24:01.608104 4954 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 17:24:01 crc kubenswrapper[4954]: E1209 17:24:01.608197 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert podName:074bde7a-88ab-45ac-8a29-09f49834e4c8 nodeName:}" failed. No retries permitted until 2025-12-09 17:24:05.608173566 +0000 UTC m=+1641.996347386 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fzcn6j" (UID: "074bde7a-88ab-45ac-8a29-09f49834e4c8") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 17:24:01 crc kubenswrapper[4954]: I1209 17:24:01.812908 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:01 crc kubenswrapper[4954]: I1209 17:24:01.812977 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:01 crc kubenswrapper[4954]: E1209 17:24:01.813160 4954 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 17:24:01 crc kubenswrapper[4954]: E1209 17:24:01.813249 4954 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 17:24:01 crc kubenswrapper[4954]: E1209 17:24:01.813284 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:24:05.813255456 +0000 UTC m=+1642.201429466 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "webhook-server-cert" not found Dec 09 17:24:01 crc kubenswrapper[4954]: E1209 17:24:01.813411 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:24:05.81336531 +0000 UTC m=+1642.201539290 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "metrics-server-cert" not found Dec 09 17:24:04 crc kubenswrapper[4954]: I1209 17:24:04.576984 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:24:04 crc kubenswrapper[4954]: E1209 17:24:04.577125 4954 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 17:24:04 crc kubenswrapper[4954]: E1209 17:24:04.577234 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert podName:2ec773b3-dd1d-4662-9ebb-4be2993f76a2 nodeName:}" failed. No retries permitted until 2025-12-09 17:24:12.577208963 +0000 UTC m=+1648.965382783 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert") pod "infra-operator-controller-manager-78d48bff9d-cdpd9" (UID: "2ec773b3-dd1d-4662-9ebb-4be2993f76a2") : secret "infra-operator-webhook-server-cert" not found Dec 09 17:24:05 crc kubenswrapper[4954]: I1209 17:24:05.700303 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:24:05 crc kubenswrapper[4954]: E1209 17:24:05.701431 4954 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 17:24:05 crc kubenswrapper[4954]: E1209 17:24:05.701541 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert podName:074bde7a-88ab-45ac-8a29-09f49834e4c8 nodeName:}" failed. No retries permitted until 2025-12-09 17:24:13.7015141 +0000 UTC m=+1650.089687920 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fzcn6j" (UID: "074bde7a-88ab-45ac-8a29-09f49834e4c8") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 17:24:05 crc kubenswrapper[4954]: I1209 17:24:05.904953 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:05 crc kubenswrapper[4954]: I1209 17:24:05.905019 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:05 crc kubenswrapper[4954]: E1209 17:24:05.905182 4954 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 17:24:05 crc kubenswrapper[4954]: E1209 17:24:05.905276 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:24:13.905250688 +0000 UTC m=+1650.293424508 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "webhook-server-cert" not found Dec 09 17:24:05 crc kubenswrapper[4954]: E1209 17:24:05.905383 4954 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 17:24:05 crc kubenswrapper[4954]: E1209 17:24:05.905495 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:24:13.905469035 +0000 UTC m=+1650.293642855 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "metrics-server-cert" not found Dec 09 17:24:10 crc kubenswrapper[4954]: E1209 17:24:10.759080 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 09 17:24:10 crc kubenswrapper[4954]: E1209 17:24:10.759792 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-44rpg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-4pcjs_openstack-operators(e57eccbb-b484-45ad-a474-292da5bb6547): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:10 crc kubenswrapper[4954]: E1209 17:24:10.868265 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.23:5001/openstack-k8s-operators/telemetry-operator:c4794e7165126ca78a1af546bb4ba50c90b5c4e1" Dec 09 17:24:10 crc kubenswrapper[4954]: E1209 17:24:10.868345 4954 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.23:5001/openstack-k8s-operators/telemetry-operator:c4794e7165126ca78a1af546bb4ba50c90b5c4e1" Dec 09 17:24:10 crc kubenswrapper[4954]: E1209 17:24:10.868528 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.23:5001/openstack-k8s-operators/telemetry-operator:c4794e7165126ca78a1af546bb4ba50c90b5c4e1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m9ht6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-796785f986-j59bx_openstack-operators(93405692-e3d5-4a81-97db-170d70b6cbc3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:11 crc kubenswrapper[4954]: E1209 17:24:11.893188 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 09 17:24:11 crc kubenswrapper[4954]: E1209 17:24:11.894174 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n7x6v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-n7kmf_openstack-operators(281087c4-abfa-4537-80b5-130b03885954): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:12 crc kubenswrapper[4954]: I1209 17:24:12.583294 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:24:12 crc kubenswrapper[4954]: E1209 17:24:12.583528 4954 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 17:24:12 crc kubenswrapper[4954]: E1209 17:24:12.584103 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert podName:2ec773b3-dd1d-4662-9ebb-4be2993f76a2 nodeName:}" failed. No retries permitted until 2025-12-09 17:24:28.584076092 +0000 UTC m=+1664.972249913 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert") pod "infra-operator-controller-manager-78d48bff9d-cdpd9" (UID: "2ec773b3-dd1d-4662-9ebb-4be2993f76a2") : secret "infra-operator-webhook-server-cert" not found Dec 09 17:24:13 crc kubenswrapper[4954]: E1209 17:24:13.472151 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 09 17:24:13 crc kubenswrapper[4954]: E1209 17:24:13.472458 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-njs4z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-mz7gh_openstack-operators(dc3d4dbd-0b2e-489e-b686-0195c372a917): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:13 crc kubenswrapper[4954]: I1209 17:24:13.706033 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:24:13 crc kubenswrapper[4954]: I1209 17:24:13.714843 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/074bde7a-88ab-45ac-8a29-09f49834e4c8-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fzcn6j\" (UID: \"074bde7a-88ab-45ac-8a29-09f49834e4c8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:24:13 crc kubenswrapper[4954]: I1209 17:24:13.755742 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:24:13 crc kubenswrapper[4954]: I1209 17:24:13.755822 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:24:13 crc kubenswrapper[4954]: I1209 17:24:13.909712 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:13 crc kubenswrapper[4954]: I1209 17:24:13.909792 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:13 crc kubenswrapper[4954]: E1209 17:24:13.909942 4954 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 17:24:13 crc kubenswrapper[4954]: E1209 17:24:13.910046 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:24:29.910022842 +0000 UTC m=+1666.298196662 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "metrics-server-cert" not found Dec 09 17:24:13 crc kubenswrapper[4954]: E1209 17:24:13.909942 4954 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 17:24:13 crc kubenswrapper[4954]: E1209 17:24:13.910098 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs podName:4ec17be0-419e-48e3-a00f-58741723b85d nodeName:}" failed. No retries permitted until 2025-12-09 17:24:29.910084693 +0000 UTC m=+1666.298258523 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs") pod "openstack-operator-controller-manager-866b78c4d6-m6k5w" (UID: "4ec17be0-419e-48e3-a00f-58741723b85d") : secret "webhook-server-cert" not found Dec 09 17:24:13 crc kubenswrapper[4954]: I1209 17:24:13.924805 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:24:14 crc kubenswrapper[4954]: I1209 17:24:14.251552 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" event={"ID":"3c0851fe-4209-4e7f-9622-dba385e6bc78","Type":"ContainerStarted","Data":"05d37a25afad7dc3e1244ac49ebb1224586870d5ff6764198c25a3f94ac56bfa"} Dec 09 17:24:14 crc kubenswrapper[4954]: I1209 17:24:14.263213 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" event={"ID":"54373352-1448-443c-82eb-4bd28627acdf","Type":"ContainerStarted","Data":"951d3c4698dc2f41b501b22b8c4b6abef2c519b755de102e7769a3243dacf2d8"} Dec 09 17:24:14 crc kubenswrapper[4954]: I1209 17:24:14.265433 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" event={"ID":"1f87c2e5-6b6b-4f67-b2f6-caf321fbe844","Type":"ContainerStarted","Data":"b5d2ce536574df943e068b0d61b33974cfd6f6f07113f11b5547bdcba7581ef3"} Dec 09 17:24:14 crc kubenswrapper[4954]: I1209 17:24:14.267467 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" event={"ID":"917a4c41-4cec-49ef-95f7-6cda64da9c44","Type":"ContainerStarted","Data":"9784426a9af380ad9ed03830937ac594b752b6bbc6a5d4ac6a1d72eeb16d1c98"} Dec 09 17:24:14 crc kubenswrapper[4954]: I1209 17:24:14.269537 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" event={"ID":"2b64512f-8bdc-42b5-9269-5ea7cd1eabf1","Type":"ContainerStarted","Data":"215eb48ae57ed379f6432f1a9883749b21d33e0a4fe00abab06520267fa36cea"} Dec 09 17:24:14 crc kubenswrapper[4954]: I1209 17:24:14.271047 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" event={"ID":"74b54335-7821-4cdc-a994-1dbe6c1653c9","Type":"ContainerStarted","Data":"7bf51961baa13bd23fdc8c4339d053cb7e1dba20dc9fea3ad426a1d85fa0cf47"} Dec 09 17:24:14 crc kubenswrapper[4954]: I1209 17:24:14.273013 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" event={"ID":"3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8","Type":"ContainerStarted","Data":"c01b2e5b65a3413ab13ce219b17b14921c7d2d733b2bedf6be5c677c1feb4672"} Dec 09 17:24:14 crc kubenswrapper[4954]: I1209 17:24:14.276236 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" event={"ID":"41401dcd-a6d4-4e4b-93e9-bea5977d7e90","Type":"ContainerStarted","Data":"09c3f89d878c84b4a82645c5028e3dc992383e7d697337a41e3629a21a92681f"} Dec 09 17:24:15 crc kubenswrapper[4954]: I1209 17:24:15.307450 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" event={"ID":"8f82f473-cbee-4499-89b6-e6fbc32405aa","Type":"ContainerStarted","Data":"1dec747d48eb97504bccb1bc2a2d2048feea8c7d35f71fe3175edd6d859e3180"} Dec 09 17:24:15 crc kubenswrapper[4954]: I1209 17:24:15.320254 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" event={"ID":"37d5d0b5-1e21-4b84-bd10-f4f738ef37a1","Type":"ContainerStarted","Data":"af291edbcb707e4a66b5d148be4b3a7fa40ef4b796ccc179a411cd1381ae35cf"} Dec 09 17:24:15 crc kubenswrapper[4954]: I1209 17:24:15.322429 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" event={"ID":"410c21d8-1970-411a-8704-db44a36c74cd","Type":"ContainerStarted","Data":"7fdb3c77e31b32a0538ca8f0ac41e8e322b681918f7d173452d93f38afc82f13"} Dec 09 17:24:15 crc kubenswrapper[4954]: I1209 17:24:15.342173 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" event={"ID":"6000ab00-d760-4e35-9ef6-47b878a8ef96","Type":"ContainerStarted","Data":"57a5754fe06d0e53ebc281bd64d976b3b3289090eef37c2b9248941c53c448de"} Dec 09 17:24:16 crc kubenswrapper[4954]: I1209 17:24:16.294822 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j"] Dec 09 17:24:22 crc kubenswrapper[4954]: I1209 17:24:22.941849 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-t56bj" podUID="b6f3c78e-58d6-4bba-a14e-85e3acd23e24" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 17:24:25 crc kubenswrapper[4954]: I1209 17:24:25.470035 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" event={"ID":"074bde7a-88ab-45ac-8a29-09f49834e4c8","Type":"ContainerStarted","Data":"88634966d117a531960bd4ac627ead787e26c099c871bf60ac0dbe641315116f"} Dec 09 17:24:28 crc kubenswrapper[4954]: I1209 17:24:28.638715 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:24:28 crc kubenswrapper[4954]: I1209 17:24:28.650907 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ec773b3-dd1d-4662-9ebb-4be2993f76a2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-cdpd9\" (UID: \"2ec773b3-dd1d-4662-9ebb-4be2993f76a2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:24:28 crc kubenswrapper[4954]: I1209 17:24:28.713878 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:24:29 crc kubenswrapper[4954]: E1209 17:24:29.954653 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 17:24:29 crc kubenswrapper[4954]: E1209 17:24:29.954963 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7qmxw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-hlx7x_openstack-operators(2b64512f-8bdc-42b5-9269-5ea7cd1eabf1): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 09 17:24:29 crc kubenswrapper[4954]: E1209 17:24:29.956539 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" podUID="2b64512f-8bdc-42b5-9269-5ea7cd1eabf1" Dec 09 17:24:29 crc kubenswrapper[4954]: E1209 17:24:29.962467 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 17:24:29 crc kubenswrapper[4954]: E1209 17:24:29.962796 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m9ht6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-796785f986-j59bx_openstack-operators(93405692-e3d5-4a81-97db-170d70b6cbc3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:29 crc kubenswrapper[4954]: E1209 17:24:29.966714 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" podUID="93405692-e3d5-4a81-97db-170d70b6cbc3" Dec 09 17:24:29 crc kubenswrapper[4954]: I1209 17:24:29.971721 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:29 crc kubenswrapper[4954]: I1209 17:24:29.971793 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:29 crc kubenswrapper[4954]: I1209 17:24:29.981187 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-webhook-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:29 crc kubenswrapper[4954]: I1209 17:24:29.982057 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ec17be0-419e-48e3-a00f-58741723b85d-metrics-certs\") pod \"openstack-operator-controller-manager-866b78c4d6-m6k5w\" (UID: \"4ec17be0-419e-48e3-a00f-58741723b85d\") " pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:30 crc kubenswrapper[4954]: I1209 17:24:30.170069 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:30 crc kubenswrapper[4954]: I1209 17:24:30.514663 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" Dec 09 17:24:30 crc kubenswrapper[4954]: E1209 17:24:30.517187 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" podUID="2b64512f-8bdc-42b5-9269-5ea7cd1eabf1" Dec 09 17:24:30 crc kubenswrapper[4954]: I1209 17:24:30.522422 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" Dec 09 17:24:31 crc kubenswrapper[4954]: E1209 17:24:31.227911 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 17:24:31 crc kubenswrapper[4954]: E1209 17:24:31.228146 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n7x6v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-n7kmf_openstack-operators(281087c4-abfa-4537-80b5-130b03885954): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:31 crc kubenswrapper[4954]: E1209 17:24:31.230161 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" podUID="281087c4-abfa-4537-80b5-130b03885954" Dec 09 17:24:31 crc kubenswrapper[4954]: E1209 17:24:31.243988 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 17:24:31 crc kubenswrapper[4954]: E1209 17:24:31.244287 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-44rpg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-4pcjs_openstack-operators(e57eccbb-b484-45ad-a474-292da5bb6547): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:31 crc kubenswrapper[4954]: E1209 17:24:31.245715 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" podUID="e57eccbb-b484-45ad-a474-292da5bb6547" Dec 09 17:24:31 crc kubenswrapper[4954]: E1209 17:24:31.258356 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 17:24:31 crc kubenswrapper[4954]: E1209 17:24:31.258547 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-njs4z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-mz7gh_openstack-operators(dc3d4dbd-0b2e-489e-b686-0195c372a917): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:31 crc kubenswrapper[4954]: E1209 17:24:31.259853 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" podUID="dc3d4dbd-0b2e-489e-b686-0195c372a917" Dec 09 17:24:31 crc kubenswrapper[4954]: E1209 17:24:31.527045 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" podUID="2b64512f-8bdc-42b5-9269-5ea7cd1eabf1" Dec 09 17:24:32 crc kubenswrapper[4954]: E1209 17:24:32.533727 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" podUID="2b64512f-8bdc-42b5-9269-5ea7cd1eabf1" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.858946 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.859158 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6gxmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-28bcg_openstack-operators(74b54335-7821-4cdc-a994-1dbe6c1653c9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.859183 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.859394 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gxwdd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-8ltxl_openstack-operators(3c0851fe-4209-4e7f-9622-dba385e6bc78): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.860587 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" podUID="74b54335-7821-4cdc-a994-1dbe6c1653c9" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.860651 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" podUID="3c0851fe-4209-4e7f-9622-dba385e6bc78" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.886265 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.886401 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hkqkh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-8857b_openstack-operators(41401dcd-a6d4-4e4b-93e9-bea5977d7e90): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.887867 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" podUID="41401dcd-a6d4-4e4b-93e9-bea5977d7e90" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.891362 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.891448 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lrw8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-k5vlj_openstack-operators(3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:24:33 crc kubenswrapper[4954]: E1209 17:24:33.892614 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" podUID="3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8" Dec 09 17:24:34 crc kubenswrapper[4954]: I1209 17:24:34.552945 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" Dec 09 17:24:34 crc kubenswrapper[4954]: I1209 17:24:34.553432 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" Dec 09 17:24:34 crc kubenswrapper[4954]: I1209 17:24:34.553454 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" Dec 09 17:24:34 crc kubenswrapper[4954]: I1209 17:24:34.558518 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" Dec 09 17:24:34 crc kubenswrapper[4954]: I1209 17:24:34.558728 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" Dec 09 17:24:34 crc kubenswrapper[4954]: I1209 17:24:34.559284 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" Dec 09 17:24:35 crc kubenswrapper[4954]: I1209 17:24:35.432972 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9"] Dec 09 17:24:35 crc kubenswrapper[4954]: W1209 17:24:35.482426 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ec773b3_dd1d_4662_9ebb_4be2993f76a2.slice/crio-456d8b591d3d1a1281394173a400d35def2b08ee6b12f52b214ffcd8a3be3858 WatchSource:0}: Error finding container 456d8b591d3d1a1281394173a400d35def2b08ee6b12f52b214ffcd8a3be3858: Status 404 returned error can't find the container with id 456d8b591d3d1a1281394173a400d35def2b08ee6b12f52b214ffcd8a3be3858 Dec 09 17:24:35 crc kubenswrapper[4954]: I1209 17:24:35.566139 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w"] Dec 09 17:24:35 crc kubenswrapper[4954]: I1209 17:24:35.568888 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" event={"ID":"ce9db630-dee1-4c9f-b90e-31d263b8a213","Type":"ContainerStarted","Data":"5fd396817a4a62321006aa33df7aa6fa835546e1b470048c6d6e220c823da4c7"} Dec 09 17:24:35 crc kubenswrapper[4954]: I1209 17:24:35.570311 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" event={"ID":"2ba970e1-21a2-477e-947f-89de7bfbb6a6","Type":"ContainerStarted","Data":"9c322325551d1ed135e1e85eb8d45a715eef9f311e6cae950f911629f1973918"} Dec 09 17:24:35 crc kubenswrapper[4954]: I1209 17:24:35.571788 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" event={"ID":"dc3d4dbd-0b2e-489e-b686-0195c372a917","Type":"ContainerStarted","Data":"c0c319d20c96ecc3f787160453b843ceba16e89dde6f50ae61011f327fe4347b"} Dec 09 17:24:35 crc kubenswrapper[4954]: I1209 17:24:35.573006 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" event={"ID":"2ec773b3-dd1d-4662-9ebb-4be2993f76a2","Type":"ContainerStarted","Data":"456d8b591d3d1a1281394173a400d35def2b08ee6b12f52b214ffcd8a3be3858"} Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.614887 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" event={"ID":"e84bf071-0cf4-4e92-9060-6cc7d30984db","Type":"ContainerStarted","Data":"99abc79ddfb063dd6038d318da69d9adfde11a593d4286887d994b7ed94bca49"} Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.616871 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" event={"ID":"281087c4-abfa-4537-80b5-130b03885954","Type":"ContainerStarted","Data":"36dcaf7d6467b91756ac196ed17e264c95969b8fa797e5e036cb1ddcf49c57a2"} Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.659551 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" event={"ID":"41401dcd-a6d4-4e4b-93e9-bea5977d7e90","Type":"ContainerStarted","Data":"bfaf2dfe415993fe31747983a0bd04f5f0577009edae6dae9ee264cdcf753a7d"} Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.661891 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.676380 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.677752 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" event={"ID":"4ec17be0-419e-48e3-a00f-58741723b85d","Type":"ContainerStarted","Data":"32d709ef9380cb9ec030a81bf3617c48023324064a83ff3c2ff0f98f4f969554"} Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.696013 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" event={"ID":"074bde7a-88ab-45ac-8a29-09f49834e4c8","Type":"ContainerStarted","Data":"81e2028658ade4dea6df4ae75a94c73d6c3347dba5a727e7889248436dd9bb65"} Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.700400 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-8857b" podStartSLOduration=26.206583909 podStartE2EDuration="40.700379554s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:58.946931905 +0000 UTC m=+1635.335105725" lastFinishedPulling="2025-12-09 17:24:13.44072755 +0000 UTC m=+1649.828901370" observedRunningTime="2025-12-09 17:24:36.695001905 +0000 UTC m=+1673.083175725" watchObservedRunningTime="2025-12-09 17:24:36.700379554 +0000 UTC m=+1673.088553374" Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.708442 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" event={"ID":"e57eccbb-b484-45ad-a474-292da5bb6547","Type":"ContainerStarted","Data":"1573d395def9fd4ec2133f7b3ff1d82b00e3d5a3f373be81cf195da636cb6b37"} Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.711289 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" event={"ID":"93405692-e3d5-4a81-97db-170d70b6cbc3","Type":"ContainerStarted","Data":"367b76b203dac6a10e76094b7929e6bbf317d1758c5e84d2c1662f94c84f1144"} Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.731903 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" event={"ID":"37d5d0b5-1e21-4b84-bd10-f4f738ef37a1","Type":"ContainerStarted","Data":"167f7b4969e590a47df180aa2c80ec5aeb06ac5d1ff53cc6f0558405c3c56f8e"} Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.767896 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" event={"ID":"410c21d8-1970-411a-8704-db44a36c74cd","Type":"ContainerStarted","Data":"b53f7367383d4e9e386032f216c6e179326cfff30d1985152198485a9867728f"} Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.772857 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.799929 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" Dec 09 17:24:36 crc kubenswrapper[4954]: I1209 17:24:36.826808 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nvdgc" podStartSLOduration=5.19881771 podStartE2EDuration="40.82678214s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.69075835 +0000 UTC m=+1636.078932170" lastFinishedPulling="2025-12-09 17:24:35.31872278 +0000 UTC m=+1671.706896600" observedRunningTime="2025-12-09 17:24:36.810065037 +0000 UTC m=+1673.198238857" watchObservedRunningTime="2025-12-09 17:24:36.82678214 +0000 UTC m=+1673.214955960" Dec 09 17:24:37 crc kubenswrapper[4954]: I1209 17:24:37.814956 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" event={"ID":"3c0851fe-4209-4e7f-9622-dba385e6bc78","Type":"ContainerStarted","Data":"f5e68e8ea0e2f4fbcd146541eef789a0b7afb3b2b7edbfea6aed657fa622b88a"} Dec 09 17:24:37 crc kubenswrapper[4954]: I1209 17:24:37.818531 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" event={"ID":"54373352-1448-443c-82eb-4bd28627acdf","Type":"ContainerStarted","Data":"d55289fed8df50d907aa752c5c03c7937e9c58b95914f0d9e110583842702edb"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.826386 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" event={"ID":"2ba970e1-21a2-477e-947f-89de7bfbb6a6","Type":"ContainerStarted","Data":"0698ca547a6b817f169ed3056cf19cc5c9048c067a8915cc1089dde6f431ff0b"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.827958 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.834429 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" event={"ID":"1fc4c8d7-5079-45d0-8a68-d2523e7b981f","Type":"ContainerStarted","Data":"20f44db85024bafe57df1fb991901a0de83d5454089d0bb4d71769bd78eb1649"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.840024 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" event={"ID":"e84bf071-0cf4-4e92-9060-6cc7d30984db","Type":"ContainerStarted","Data":"142ce51ef1a8a2e8e09afe5c674a012245fbd89681db6c9e435cd3d2c93c6006"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.840188 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.847379 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" event={"ID":"dc3d4dbd-0b2e-489e-b686-0195c372a917","Type":"ContainerStarted","Data":"1b9f556a91c99e143f204f5e1bdd1a49c674d6b13daa12159b0f44445a2b8e30"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.848421 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.855605 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" podStartSLOduration=8.674221441 podStartE2EDuration="42.855562433s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.703561642 +0000 UTC m=+1636.091735462" lastFinishedPulling="2025-12-09 17:24:33.884902634 +0000 UTC m=+1670.273076454" observedRunningTime="2025-12-09 17:24:38.853098966 +0000 UTC m=+1675.241272796" watchObservedRunningTime="2025-12-09 17:24:38.855562433 +0000 UTC m=+1675.243736253" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.876250 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" podStartSLOduration=7.5820391879999995 podStartE2EDuration="42.876227419s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.69711879 +0000 UTC m=+1636.085292610" lastFinishedPulling="2025-12-09 17:24:34.991307021 +0000 UTC m=+1671.379480841" observedRunningTime="2025-12-09 17:24:38.875529578 +0000 UTC m=+1675.263703398" watchObservedRunningTime="2025-12-09 17:24:38.876227419 +0000 UTC m=+1675.264401239" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.880129 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" event={"ID":"074bde7a-88ab-45ac-8a29-09f49834e4c8","Type":"ContainerStarted","Data":"4f09a37f0e85a9a564eef918dd79a0798166cf05839863ed15d847344d71910c"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.881182 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.889755 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" event={"ID":"8f82f473-cbee-4499-89b6-e6fbc32405aa","Type":"ContainerStarted","Data":"b914a6f1e02b1e7aaf3d7d20f6a0f79d1aa396d387e3b34dacc50e264450889a"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.891254 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.894023 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.901936 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rcwxm" podStartSLOduration=6.486088715 podStartE2EDuration="41.901920724s" podCreationTimestamp="2025-12-09 17:23:57 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.752582566 +0000 UTC m=+1636.140756386" lastFinishedPulling="2025-12-09 17:24:35.168414575 +0000 UTC m=+1671.556588395" observedRunningTime="2025-12-09 17:24:38.894659877 +0000 UTC m=+1675.282833697" watchObservedRunningTime="2025-12-09 17:24:38.901920724 +0000 UTC m=+1675.290094544" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.906674 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" event={"ID":"1f87c2e5-6b6b-4f67-b2f6-caf321fbe844","Type":"ContainerStarted","Data":"9e23a8c11c5a128b9f73a6d7fd05343c60fa837893984d59fb4aa136d06f61c1"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.907458 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.909679 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.913099 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" event={"ID":"917a4c41-4cec-49ef-95f7-6cda64da9c44","Type":"ContainerStarted","Data":"d9658f69c6cdd49636cc270d20dad87e03f36f391186ccb469ed220181406959"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.914340 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.918216 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.918798 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" event={"ID":"93405692-e3d5-4a81-97db-170d70b6cbc3","Type":"ContainerStarted","Data":"524eae5a2ced3744765b462a231dc319343ddef3b930c7d506bef77e73586ad9"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.920476 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.922852 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" event={"ID":"e57eccbb-b484-45ad-a474-292da5bb6547","Type":"ContainerStarted","Data":"736a40ccc75573d75881bda78611018bc50996732771ae439e4b7af0b35a297b"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.923899 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.938631 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" event={"ID":"74b54335-7821-4cdc-a994-1dbe6c1653c9","Type":"ContainerStarted","Data":"c32b145bdb2ff8af4ed78c945832bbc6386b30934782847c7d6c3c6527951d55"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.941938 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" event={"ID":"3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8","Type":"ContainerStarted","Data":"0942ec23dac663522a7dea45fc52fee971d8f50837b056b57f3d86b5ef528514"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.954963 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" event={"ID":"4ec17be0-419e-48e3-a00f-58741723b85d","Type":"ContainerStarted","Data":"27544a5aae058807734f46de2028d5dd7f3d842c80ecae99c7c1b4bb3ab120b7"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.956038 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.959378 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-4rrqn" podStartSLOduration=6.623378573 podStartE2EDuration="41.959354482s" podCreationTimestamp="2025-12-09 17:23:57 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.738895727 +0000 UTC m=+1636.127069547" lastFinishedPulling="2025-12-09 17:24:35.074871636 +0000 UTC m=+1671.463045456" observedRunningTime="2025-12-09 17:24:38.955077648 +0000 UTC m=+1675.343251478" watchObservedRunningTime="2025-12-09 17:24:38.959354482 +0000 UTC m=+1675.347528302" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.962014 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" podStartSLOduration=7.359280455 podStartE2EDuration="42.961999715s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.519936702 +0000 UTC m=+1635.908110532" lastFinishedPulling="2025-12-09 17:24:35.122655972 +0000 UTC m=+1671.510829792" observedRunningTime="2025-12-09 17:24:38.928441125 +0000 UTC m=+1675.316614965" watchObservedRunningTime="2025-12-09 17:24:38.961999715 +0000 UTC m=+1675.350173535" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.965834 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" event={"ID":"6000ab00-d760-4e35-9ef6-47b878a8ef96","Type":"ContainerStarted","Data":"2b73196ead0906c9c4d8fec208e99bb0c1b954bedbbd3a1d2dacd9c1a5bfa9a2"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.966388 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.969470 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.970334 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" event={"ID":"ce9db630-dee1-4c9f-b90e-31d263b8a213","Type":"ContainerStarted","Data":"97e3730988cf634a4900878552e74e36cfc60abbad8ecf434e4050c549eabf4e"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.972250 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.976180 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" event={"ID":"281087c4-abfa-4537-80b5-130b03885954","Type":"ContainerStarted","Data":"d4be8b02a280586e4257adc0a8b0118d64cb24e328ee50a307653391d4953fe2"} Dec 09 17:24:38 crc kubenswrapper[4954]: I1209 17:24:38.977529 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.001058 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.033623 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" podStartSLOduration=6.17947187 podStartE2EDuration="43.033577705s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:58.268531247 +0000 UTC m=+1634.656705067" lastFinishedPulling="2025-12-09 17:24:35.122637082 +0000 UTC m=+1671.510810902" observedRunningTime="2025-12-09 17:24:38.992551451 +0000 UTC m=+1675.380725301" watchObservedRunningTime="2025-12-09 17:24:39.033577705 +0000 UTC m=+1675.421751535" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.065032 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" podStartSLOduration=33.160843695 podStartE2EDuration="43.06500513s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:24:25.23358181 +0000 UTC m=+1661.621755630" lastFinishedPulling="2025-12-09 17:24:35.137743245 +0000 UTC m=+1671.525917065" observedRunningTime="2025-12-09 17:24:39.026318599 +0000 UTC m=+1675.414492429" watchObservedRunningTime="2025-12-09 17:24:39.06500513 +0000 UTC m=+1675.453178950" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.077820 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-8mpbt" podStartSLOduration=6.658605849 podStartE2EDuration="43.077691286s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:58.939372658 +0000 UTC m=+1635.327546478" lastFinishedPulling="2025-12-09 17:24:35.358458095 +0000 UTC m=+1671.746631915" observedRunningTime="2025-12-09 17:24:39.056000587 +0000 UTC m=+1675.444174427" watchObservedRunningTime="2025-12-09 17:24:39.077691286 +0000 UTC m=+1675.465865106" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.095337 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-pfvmf" podStartSLOduration=7.277063641 podStartE2EDuration="43.095315868s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.477729401 +0000 UTC m=+1635.865903221" lastFinishedPulling="2025-12-09 17:24:35.295981628 +0000 UTC m=+1671.684155448" observedRunningTime="2025-12-09 17:24:39.085703807 +0000 UTC m=+1675.473877627" watchObservedRunningTime="2025-12-09 17:24:39.095315868 +0000 UTC m=+1675.483489678" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.121001 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" podStartSLOduration=7.494794228 podStartE2EDuration="43.120976492s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.496451308 +0000 UTC m=+1635.884625128" lastFinishedPulling="2025-12-09 17:24:35.122633572 +0000 UTC m=+1671.510807392" observedRunningTime="2025-12-09 17:24:39.113540859 +0000 UTC m=+1675.501714689" watchObservedRunningTime="2025-12-09 17:24:39.120976492 +0000 UTC m=+1675.509150312" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.164519 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-28bcg" podStartSLOduration=29.148088054 podStartE2EDuration="43.164494034s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.425891879 +0000 UTC m=+1635.814065689" lastFinishedPulling="2025-12-09 17:24:13.442297849 +0000 UTC m=+1649.830471669" observedRunningTime="2025-12-09 17:24:39.144201999 +0000 UTC m=+1675.532375819" watchObservedRunningTime="2025-12-09 17:24:39.164494034 +0000 UTC m=+1675.552667854" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.216723 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" podStartSLOduration=6.394500683 podStartE2EDuration="43.216694288s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:58.356935825 +0000 UTC m=+1634.745109645" lastFinishedPulling="2025-12-09 17:24:35.17912942 +0000 UTC m=+1671.567303250" observedRunningTime="2025-12-09 17:24:39.207925074 +0000 UTC m=+1675.596098894" watchObservedRunningTime="2025-12-09 17:24:39.216694288 +0000 UTC m=+1675.604868118" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.253021 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" podStartSLOduration=7.909599454 podStartE2EDuration="43.252985585s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.701051913 +0000 UTC m=+1636.089225733" lastFinishedPulling="2025-12-09 17:24:35.044438044 +0000 UTC m=+1671.432611864" observedRunningTime="2025-12-09 17:24:39.243077424 +0000 UTC m=+1675.631251244" watchObservedRunningTime="2025-12-09 17:24:39.252985585 +0000 UTC m=+1675.641159405" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.274557 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-qsdpz" podStartSLOduration=7.450634025 podStartE2EDuration="43.274529979s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.478550307 +0000 UTC m=+1635.866724127" lastFinishedPulling="2025-12-09 17:24:35.302446261 +0000 UTC m=+1671.690620081" observedRunningTime="2025-12-09 17:24:39.268032405 +0000 UTC m=+1675.656206235" watchObservedRunningTime="2025-12-09 17:24:39.274529979 +0000 UTC m=+1675.662703809" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.332196 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-k5vlj" podStartSLOduration=28.805034774 podStartE2EDuration="43.332164873s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:58.918446193 +0000 UTC m=+1635.306620013" lastFinishedPulling="2025-12-09 17:24:13.445576292 +0000 UTC m=+1649.833750112" observedRunningTime="2025-12-09 17:24:39.297349153 +0000 UTC m=+1675.685522983" watchObservedRunningTime="2025-12-09 17:24:39.332164873 +0000 UTC m=+1675.720338693" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.375244 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" podStartSLOduration=7.664993896 podStartE2EDuration="43.375218051s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.412410947 +0000 UTC m=+1635.800584767" lastFinishedPulling="2025-12-09 17:24:35.122635102 +0000 UTC m=+1671.510808922" observedRunningTime="2025-12-09 17:24:39.321250481 +0000 UTC m=+1675.709424301" watchObservedRunningTime="2025-12-09 17:24:39.375218051 +0000 UTC m=+1675.763391871" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.407053 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-g6ggb" podStartSLOduration=7.339393223 podStartE2EDuration="43.407017367s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.252193761 +0000 UTC m=+1635.640367581" lastFinishedPulling="2025-12-09 17:24:35.319817895 +0000 UTC m=+1671.707991725" observedRunningTime="2025-12-09 17:24:39.368332526 +0000 UTC m=+1675.756506356" watchObservedRunningTime="2025-12-09 17:24:39.407017367 +0000 UTC m=+1675.795191197" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.425504 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" podStartSLOduration=42.425473464 podStartE2EDuration="42.425473464s" podCreationTimestamp="2025-12-09 17:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:24:39.40458007 +0000 UTC m=+1675.792753910" watchObservedRunningTime="2025-12-09 17:24:39.425473464 +0000 UTC m=+1675.813647294" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.437652 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8ltxl" podStartSLOduration=29.690314128 podStartE2EDuration="43.437630175s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.691008648 +0000 UTC m=+1636.079182468" lastFinishedPulling="2025-12-09 17:24:13.438324695 +0000 UTC m=+1649.826498515" observedRunningTime="2025-12-09 17:24:39.435106486 +0000 UTC m=+1675.823280316" watchObservedRunningTime="2025-12-09 17:24:39.437630175 +0000 UTC m=+1675.825804005" Dec 09 17:24:39 crc kubenswrapper[4954]: I1209 17:24:39.990276 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" Dec 09 17:24:40 crc kubenswrapper[4954]: I1209 17:24:40.999073 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" event={"ID":"2ec773b3-dd1d-4662-9ebb-4be2993f76a2","Type":"ContainerStarted","Data":"cab1e8d1f2ea15f35cab0106cda47d8a36e229589bf5303bda6eaae4fcf7154f"} Dec 09 17:24:40 crc kubenswrapper[4954]: I1209 17:24:40.999479 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" event={"ID":"2ec773b3-dd1d-4662-9ebb-4be2993f76a2","Type":"ContainerStarted","Data":"6a49a035752d2499efc10b255d538dd54d0b6682d0f5d81379768b7b80eaceb2"} Dec 09 17:24:41 crc kubenswrapper[4954]: I1209 17:24:41.004187 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-96qpl" Dec 09 17:24:41 crc kubenswrapper[4954]: I1209 17:24:41.004234 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-5jjz9" Dec 09 17:24:41 crc kubenswrapper[4954]: I1209 17:24:41.004577 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4pcjs" Dec 09 17:24:41 crc kubenswrapper[4954]: I1209 17:24:41.004742 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-n7kmf" Dec 09 17:24:41 crc kubenswrapper[4954]: I1209 17:24:41.005923 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-mz7gh" Dec 09 17:24:41 crc kubenswrapper[4954]: I1209 17:24:41.008036 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fzcn6j" Dec 09 17:24:41 crc kubenswrapper[4954]: I1209 17:24:41.023393 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" podStartSLOduration=40.011752876 podStartE2EDuration="45.023371427s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:24:35.486890865 +0000 UTC m=+1671.875064685" lastFinishedPulling="2025-12-09 17:24:40.498509416 +0000 UTC m=+1676.886683236" observedRunningTime="2025-12-09 17:24:41.018754233 +0000 UTC m=+1677.406928063" watchObservedRunningTime="2025-12-09 17:24:41.023371427 +0000 UTC m=+1677.411545257" Dec 09 17:24:42 crc kubenswrapper[4954]: I1209 17:24:42.007227 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:24:43 crc kubenswrapper[4954]: I1209 17:24:43.755880 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:24:43 crc kubenswrapper[4954]: I1209 17:24:43.756291 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:24:43 crc kubenswrapper[4954]: I1209 17:24:43.756351 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:24:43 crc kubenswrapper[4954]: I1209 17:24:43.757039 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:24:43 crc kubenswrapper[4954]: I1209 17:24:43.757142 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" gracePeriod=600 Dec 09 17:24:43 crc kubenswrapper[4954]: E1209 17:24:43.881376 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:24:44 crc kubenswrapper[4954]: I1209 17:24:44.025689 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" exitCode=0 Dec 09 17:24:44 crc kubenswrapper[4954]: I1209 17:24:44.025730 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d"} Dec 09 17:24:44 crc kubenswrapper[4954]: I1209 17:24:44.025784 4954 scope.go:117] "RemoveContainer" containerID="4eb79bd8a7242b0656d839869821c630e99a7a1a245bd7247948b5ac5e84baa8" Dec 09 17:24:44 crc kubenswrapper[4954]: I1209 17:24:44.026574 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:24:44 crc kubenswrapper[4954]: E1209 17:24:44.026909 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:24:46 crc kubenswrapper[4954]: I1209 17:24:46.045333 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" event={"ID":"2b64512f-8bdc-42b5-9269-5ea7cd1eabf1","Type":"ContainerStarted","Data":"7e68e2ce3af686da0bf3994cb0d27f8f018cbc7899739a53d5cab1863d8b1c5f"} Dec 09 17:24:46 crc kubenswrapper[4954]: I1209 17:24:46.832826 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" Dec 09 17:24:46 crc kubenswrapper[4954]: I1209 17:24:46.838380 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6sfrq" Dec 09 17:24:46 crc kubenswrapper[4954]: I1209 17:24:46.862889 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hlx7x" podStartSLOduration=36.66805866 podStartE2EDuration="50.862861635s" podCreationTimestamp="2025-12-09 17:23:56 +0000 UTC" firstStartedPulling="2025-12-09 17:23:59.250266821 +0000 UTC m=+1635.638440641" lastFinishedPulling="2025-12-09 17:24:13.445069796 +0000 UTC m=+1649.833243616" observedRunningTime="2025-12-09 17:24:46.067578779 +0000 UTC m=+1682.455752609" watchObservedRunningTime="2025-12-09 17:24:46.862861635 +0000 UTC m=+1683.251035455" Dec 09 17:24:47 crc kubenswrapper[4954]: I1209 17:24:47.655873 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-76tb6" Dec 09 17:24:48 crc kubenswrapper[4954]: I1209 17:24:48.148890 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-796785f986-j59bx" Dec 09 17:24:48 crc kubenswrapper[4954]: I1209 17:24:48.720824 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-cdpd9" Dec 09 17:24:50 crc kubenswrapper[4954]: I1209 17:24:50.178648 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-866b78c4d6-m6k5w" Dec 09 17:24:58 crc kubenswrapper[4954]: I1209 17:24:58.120765 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:24:58 crc kubenswrapper[4954]: E1209 17:24:58.121446 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.340112 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tcbm7"] Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.343429 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.346101 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.347443 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.347669 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.349645 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-bxwxv" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.349654 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.361896 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tcbm7"] Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.474737 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tcbm7\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.474814 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4bkf\" (UniqueName: \"kubernetes.io/projected/a142cf6b-3845-4f0a-8e7d-0359af36b21a-kube-api-access-z4bkf\") pod \"dnsmasq-dns-78dd6ddcc-tcbm7\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.474847 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-config\") pod \"dnsmasq-dns-78dd6ddcc-tcbm7\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.576700 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tcbm7\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.577563 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4bkf\" (UniqueName: \"kubernetes.io/projected/a142cf6b-3845-4f0a-8e7d-0359af36b21a-kube-api-access-z4bkf\") pod \"dnsmasq-dns-78dd6ddcc-tcbm7\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.577719 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-config\") pod \"dnsmasq-dns-78dd6ddcc-tcbm7\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.577790 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tcbm7\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.578845 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-config\") pod \"dnsmasq-dns-78dd6ddcc-tcbm7\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.598606 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4bkf\" (UniqueName: \"kubernetes.io/projected/a142cf6b-3845-4f0a-8e7d-0359af36b21a-kube-api-access-z4bkf\") pod \"dnsmasq-dns-78dd6ddcc-tcbm7\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:07 crc kubenswrapper[4954]: I1209 17:25:07.676149 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:08 crc kubenswrapper[4954]: I1209 17:25:08.132495 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tcbm7"] Dec 09 17:25:08 crc kubenswrapper[4954]: I1209 17:25:08.244207 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" event={"ID":"a142cf6b-3845-4f0a-8e7d-0359af36b21a","Type":"ContainerStarted","Data":"92cf589e1bf97a0db61ed2fc039a6483b25af8c9fb9093dba80bc7658338636b"} Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.121532 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:25:10 crc kubenswrapper[4954]: E1209 17:25:10.122390 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.340504 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kr46n"] Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.344423 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.366319 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kr46n"] Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.469247 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59d9w\" (UniqueName: \"kubernetes.io/projected/086f5371-c7ea-49ba-bd08-2096c1e71d2c-kube-api-access-59d9w\") pod \"dnsmasq-dns-5ccc8479f9-kr46n\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.469342 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-kr46n\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.469411 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-config\") pod \"dnsmasq-dns-5ccc8479f9-kr46n\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.571133 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59d9w\" (UniqueName: \"kubernetes.io/projected/086f5371-c7ea-49ba-bd08-2096c1e71d2c-kube-api-access-59d9w\") pod \"dnsmasq-dns-5ccc8479f9-kr46n\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.571205 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-kr46n\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.571250 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-config\") pod \"dnsmasq-dns-5ccc8479f9-kr46n\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.572184 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-config\") pod \"dnsmasq-dns-5ccc8479f9-kr46n\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.573641 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-kr46n\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.626363 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59d9w\" (UniqueName: \"kubernetes.io/projected/086f5371-c7ea-49ba-bd08-2096c1e71d2c-kube-api-access-59d9w\") pod \"dnsmasq-dns-5ccc8479f9-kr46n\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.674548 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.802860 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tcbm7"] Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.845624 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-x9rc7"] Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.851486 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.862727 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-x9rc7"] Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.978120 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-x9rc7\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.978677 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp684\" (UniqueName: \"kubernetes.io/projected/62728408-1835-4e60-9d36-dac31093b2b9-kube-api-access-gp684\") pod \"dnsmasq-dns-57d769cc4f-x9rc7\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:10 crc kubenswrapper[4954]: I1209 17:25:10.978721 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-config\") pod \"dnsmasq-dns-57d769cc4f-x9rc7\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.081934 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-x9rc7\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.082057 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp684\" (UniqueName: \"kubernetes.io/projected/62728408-1835-4e60-9d36-dac31093b2b9-kube-api-access-gp684\") pod \"dnsmasq-dns-57d769cc4f-x9rc7\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.082109 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-config\") pod \"dnsmasq-dns-57d769cc4f-x9rc7\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.083229 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-config\") pod \"dnsmasq-dns-57d769cc4f-x9rc7\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.083947 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-x9rc7\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.121381 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp684\" (UniqueName: \"kubernetes.io/projected/62728408-1835-4e60-9d36-dac31093b2b9-kube-api-access-gp684\") pod \"dnsmasq-dns-57d769cc4f-x9rc7\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.201697 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.381810 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kr46n"] Dec 09 17:25:11 crc kubenswrapper[4954]: W1209 17:25:11.390820 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod086f5371_c7ea_49ba_bd08_2096c1e71d2c.slice/crio-4e93121d843d2d3bc26bac95e92f539bfd6f2d50225df499e789db0063a728d4 WatchSource:0}: Error finding container 4e93121d843d2d3bc26bac95e92f539bfd6f2d50225df499e789db0063a728d4: Status 404 returned error can't find the container with id 4e93121d843d2d3bc26bac95e92f539bfd6f2d50225df499e789db0063a728d4 Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.536859 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.539260 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.544260 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.544831 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.544881 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.545129 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.545219 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.545273 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.553550 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.556325 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-x79kt" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.696218 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.696287 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxtjv\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-kube-api-access-wxtjv\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.696313 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.696339 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.696379 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.696402 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.696422 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/28718e77-f5b5-4d67-afbc-c1157354bc47-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.696441 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/28718e77-f5b5-4d67-afbc-c1157354bc47-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.696743 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.696788 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.697085 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.714627 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-x9rc7"] Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799358 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799470 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxtjv\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-kube-api-access-wxtjv\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799506 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799544 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799616 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799650 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799674 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/28718e77-f5b5-4d67-afbc-c1157354bc47-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799701 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/28718e77-f5b5-4d67-afbc-c1157354bc47-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799733 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799768 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.799808 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.800732 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.800780 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.801052 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.801419 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.801761 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.801982 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.809148 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.809650 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/28718e77-f5b5-4d67-afbc-c1157354bc47-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.810275 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.813185 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/28718e77-f5b5-4d67-afbc-c1157354bc47-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.816967 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxtjv\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-kube-api-access-wxtjv\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.825995 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.874352 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.988208 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.990154 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.997955 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8smf5" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.998250 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.998389 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.998541 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.998669 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 09 17:25:11 crc kubenswrapper[4954]: I1209 17:25:11.999269 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.000565 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.002079 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.105816 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-config-data\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.106362 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.106443 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.106490 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.106607 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3645ed14-3d97-496e-a683-2e56ea33fec9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.106643 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.106678 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmjsq\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-kube-api-access-dmjsq\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.106780 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3645ed14-3d97-496e-a683-2e56ea33fec9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.106953 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.106996 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.107145 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209230 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3645ed14-3d97-496e-a683-2e56ea33fec9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209292 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209315 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmjsq\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-kube-api-access-dmjsq\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209342 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3645ed14-3d97-496e-a683-2e56ea33fec9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209397 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209424 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209454 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209510 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-config-data\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209535 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209587 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.209651 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.211367 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.211878 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.213090 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.214970 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.214978 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-config-data\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.215153 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.221312 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.224326 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3645ed14-3d97-496e-a683-2e56ea33fec9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.224487 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.231297 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3645ed14-3d97-496e-a683-2e56ea33fec9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.236669 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmjsq\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-kube-api-access-dmjsq\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.291162 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" event={"ID":"086f5371-c7ea-49ba-bd08-2096c1e71d2c","Type":"ContainerStarted","Data":"4e93121d843d2d3bc26bac95e92f539bfd6f2d50225df499e789db0063a728d4"} Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.293754 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" event={"ID":"62728408-1835-4e60-9d36-dac31093b2b9","Type":"ContainerStarted","Data":"d9f741b2bbcaeb18ee1a62357ff7c5061357e5714010c73c58a58c7f3043bbe2"} Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.295491 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.311912 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 17:25:12 crc kubenswrapper[4954]: I1209 17:25:12.495716 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.203254 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.205346 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.208003 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-8d748" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.212601 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.212789 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.213085 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.222163 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.225049 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.338486 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d42e4546-9029-4930-ab04-7adaab383d5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.338544 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6rwh\" (UniqueName: \"kubernetes.io/projected/d42e4546-9029-4930-ab04-7adaab383d5a-kube-api-access-k6rwh\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.338658 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d42e4546-9029-4930-ab04-7adaab383d5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.338688 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d42e4546-9029-4930-ab04-7adaab383d5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.338804 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42e4546-9029-4930-ab04-7adaab383d5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.339335 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.339473 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d42e4546-9029-4930-ab04-7adaab383d5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.339605 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d42e4546-9029-4930-ab04-7adaab383d5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.441723 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d42e4546-9029-4930-ab04-7adaab383d5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.441774 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d42e4546-9029-4930-ab04-7adaab383d5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.441830 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42e4546-9029-4930-ab04-7adaab383d5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.441921 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.441944 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d42e4546-9029-4930-ab04-7adaab383d5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.441970 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d42e4546-9029-4930-ab04-7adaab383d5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.442010 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d42e4546-9029-4930-ab04-7adaab383d5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.442027 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6rwh\" (UniqueName: \"kubernetes.io/projected/d42e4546-9029-4930-ab04-7adaab383d5a-kube-api-access-k6rwh\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.442388 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d42e4546-9029-4930-ab04-7adaab383d5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.443005 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d42e4546-9029-4930-ab04-7adaab383d5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.443081 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d42e4546-9029-4930-ab04-7adaab383d5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.443943 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.444397 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d42e4546-9029-4930-ab04-7adaab383d5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.451645 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42e4546-9029-4930-ab04-7adaab383d5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.461637 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6rwh\" (UniqueName: \"kubernetes.io/projected/d42e4546-9029-4930-ab04-7adaab383d5a-kube-api-access-k6rwh\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.472698 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.495335 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d42e4546-9029-4930-ab04-7adaab383d5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d42e4546-9029-4930-ab04-7adaab383d5a\") " pod="openstack/openstack-galera-0" Dec 09 17:25:13 crc kubenswrapper[4954]: I1209 17:25:13.553193 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.664150 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.683088 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.684377 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.692965 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.693011 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-6btnr" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.693356 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.693857 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.778194 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/197b01d4-b9ac-4dc6-ac33-4638ece0241f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.778271 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpv4t\" (UniqueName: \"kubernetes.io/projected/197b01d4-b9ac-4dc6-ac33-4638ece0241f-kube-api-access-vpv4t\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.778302 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/197b01d4-b9ac-4dc6-ac33-4638ece0241f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.778345 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197b01d4-b9ac-4dc6-ac33-4638ece0241f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.778426 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/197b01d4-b9ac-4dc6-ac33-4638ece0241f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.778488 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/197b01d4-b9ac-4dc6-ac33-4638ece0241f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.778515 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/197b01d4-b9ac-4dc6-ac33-4638ece0241f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.778543 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.882469 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/197b01d4-b9ac-4dc6-ac33-4638ece0241f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.882613 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.882680 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/197b01d4-b9ac-4dc6-ac33-4638ece0241f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.882872 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpv4t\" (UniqueName: \"kubernetes.io/projected/197b01d4-b9ac-4dc6-ac33-4638ece0241f-kube-api-access-vpv4t\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.882899 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/197b01d4-b9ac-4dc6-ac33-4638ece0241f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.882923 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197b01d4-b9ac-4dc6-ac33-4638ece0241f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.882993 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/197b01d4-b9ac-4dc6-ac33-4638ece0241f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.884212 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.884585 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/197b01d4-b9ac-4dc6-ac33-4638ece0241f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.885291 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/197b01d4-b9ac-4dc6-ac33-4638ece0241f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.885347 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/197b01d4-b9ac-4dc6-ac33-4638ece0241f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.883050 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/197b01d4-b9ac-4dc6-ac33-4638ece0241f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.886152 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/197b01d4-b9ac-4dc6-ac33-4638ece0241f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.904892 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/197b01d4-b9ac-4dc6-ac33-4638ece0241f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.908012 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpv4t\" (UniqueName: \"kubernetes.io/projected/197b01d4-b9ac-4dc6-ac33-4638ece0241f-kube-api-access-vpv4t\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.915021 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197b01d4-b9ac-4dc6-ac33-4638ece0241f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.941685 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"197b01d4-b9ac-4dc6-ac33-4638ece0241f\") " pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:14 crc kubenswrapper[4954]: I1209 17:25:14.979733 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.008533 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.021836 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-jxs8p" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.021981 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.022171 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.022543 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.073469 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.123808 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.123893 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.123986 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-config-data\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.124015 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brnwg\" (UniqueName: \"kubernetes.io/projected/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-kube-api-access-brnwg\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.124045 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-kolla-config\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.226889 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-config-data\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.226963 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brnwg\" (UniqueName: \"kubernetes.io/projected/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-kube-api-access-brnwg\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.227035 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-kolla-config\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.227202 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.227309 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.229606 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-config-data\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.229735 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-kolla-config\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.234465 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.235100 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.243374 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brnwg\" (UniqueName: \"kubernetes.io/projected/bb8de77c-60eb-4e0d-abdf-d6bc9f111031-kube-api-access-brnwg\") pod \"memcached-0\" (UID: \"bb8de77c-60eb-4e0d-abdf-d6bc9f111031\") " pod="openstack/memcached-0" Dec 09 17:25:15 crc kubenswrapper[4954]: I1209 17:25:15.341256 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 09 17:25:17 crc kubenswrapper[4954]: I1209 17:25:17.474449 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 17:25:17 crc kubenswrapper[4954]: I1209 17:25:17.476383 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 17:25:17 crc kubenswrapper[4954]: I1209 17:25:17.484101 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-cdrlq" Dec 09 17:25:17 crc kubenswrapper[4954]: I1209 17:25:17.508609 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 17:25:17 crc kubenswrapper[4954]: I1209 17:25:17.582944 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fnnt\" (UniqueName: \"kubernetes.io/projected/72a212eb-2f11-46aa-88ff-6b59b86c5b6c-kube-api-access-9fnnt\") pod \"kube-state-metrics-0\" (UID: \"72a212eb-2f11-46aa-88ff-6b59b86c5b6c\") " pod="openstack/kube-state-metrics-0" Dec 09 17:25:17 crc kubenswrapper[4954]: I1209 17:25:17.687499 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fnnt\" (UniqueName: \"kubernetes.io/projected/72a212eb-2f11-46aa-88ff-6b59b86c5b6c-kube-api-access-9fnnt\") pod \"kube-state-metrics-0\" (UID: \"72a212eb-2f11-46aa-88ff-6b59b86c5b6c\") " pod="openstack/kube-state-metrics-0" Dec 09 17:25:17 crc kubenswrapper[4954]: I1209 17:25:17.754064 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fnnt\" (UniqueName: \"kubernetes.io/projected/72a212eb-2f11-46aa-88ff-6b59b86c5b6c-kube-api-access-9fnnt\") pod \"kube-state-metrics-0\" (UID: \"72a212eb-2f11-46aa-88ff-6b59b86c5b6c\") " pod="openstack/kube-state-metrics-0" Dec 09 17:25:17 crc kubenswrapper[4954]: I1209 17:25:17.830802 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.221506 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5"] Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.222981 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.231454 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.231756 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-825hk" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.273272 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5"] Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.404063 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmp4b\" (UniqueName: \"kubernetes.io/projected/300e9948-95ee-44e4-b816-16664d86e2e2-kube-api-access-pmp4b\") pod \"observability-ui-dashboards-7d5fb4cbfb-mmts5\" (UID: \"300e9948-95ee-44e4-b816-16664d86e2e2\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.404619 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/300e9948-95ee-44e4-b816-16664d86e2e2-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-mmts5\" (UID: \"300e9948-95ee-44e4-b816-16664d86e2e2\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.506514 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmp4b\" (UniqueName: \"kubernetes.io/projected/300e9948-95ee-44e4-b816-16664d86e2e2-kube-api-access-pmp4b\") pod \"observability-ui-dashboards-7d5fb4cbfb-mmts5\" (UID: \"300e9948-95ee-44e4-b816-16664d86e2e2\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.506581 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/300e9948-95ee-44e4-b816-16664d86e2e2-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-mmts5\" (UID: \"300e9948-95ee-44e4-b816-16664d86e2e2\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.513746 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/300e9948-95ee-44e4-b816-16664d86e2e2-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-mmts5\" (UID: \"300e9948-95ee-44e4-b816-16664d86e2e2\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.581805 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmp4b\" (UniqueName: \"kubernetes.io/projected/300e9948-95ee-44e4-b816-16664d86e2e2-kube-api-access-pmp4b\") pod \"observability-ui-dashboards-7d5fb4cbfb-mmts5\" (UID: \"300e9948-95ee-44e4-b816-16664d86e2e2\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.764930 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.814120 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.820768 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.831187 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.833102 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.833290 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.833446 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-xwpb8" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.833820 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.835400 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.875031 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.916680 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-config\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.916748 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.916804 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w4n7\" (UniqueName: \"kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-kube-api-access-8w4n7\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.916857 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.916891 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f0da1b73-3249-449b-86f9-12206ce93aab-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.916912 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.916948 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f0da1b73-3249-449b-86f9-12206ce93aab-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.916996 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.984052 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-595b748fd5-4v476"] Dec 09 17:25:18 crc kubenswrapper[4954]: I1209 17:25:18.985670 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.018565 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-595b748fd5-4v476"] Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.020141 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.020298 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w4n7\" (UniqueName: \"kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-kube-api-access-8w4n7\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.020430 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.020519 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f0da1b73-3249-449b-86f9-12206ce93aab-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.020616 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.020760 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f0da1b73-3249-449b-86f9-12206ce93aab-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.020924 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.021098 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-config\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: E1209 17:25:19.022096 4954 configmap.go:193] Couldn't get configMap openstack/prometheus-metric-storage-rulefiles-0: configmap "prometheus-metric-storage-rulefiles-0" not found Dec 09 17:25:19 crc kubenswrapper[4954]: E1209 17:25:19.022221 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f0da1b73-3249-449b-86f9-12206ce93aab-prometheus-metric-storage-rulefiles-0 podName:f0da1b73-3249-449b-86f9-12206ce93aab nodeName:}" failed. No retries permitted until 2025-12-09 17:25:19.522194887 +0000 UTC m=+1715.910368697 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "prometheus-metric-storage-rulefiles-0" (UniqueName: "kubernetes.io/configmap/f0da1b73-3249-449b-86f9-12206ce93aab-prometheus-metric-storage-rulefiles-0") pod "prometheus-metric-storage-0" (UID: "f0da1b73-3249-449b-86f9-12206ce93aab") : configmap "prometheus-metric-storage-rulefiles-0" not found Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.025962 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.026875 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-config\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.038026 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.038725 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.045017 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f0da1b73-3249-449b-86f9-12206ce93aab-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.070815 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.116545 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w4n7\" (UniqueName: \"kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-kube-api-access-8w4n7\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.122118 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-service-ca\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.122169 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqwjq\" (UniqueName: \"kubernetes.io/projected/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-kube-api-access-wqwjq\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.122211 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-console-oauth-config\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.122259 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-console-config\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.122282 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-oauth-serving-cert\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.122350 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-trusted-ca-bundle\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.122382 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-console-serving-cert\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.132178 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.223695 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-service-ca\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.223800 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqwjq\" (UniqueName: \"kubernetes.io/projected/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-kube-api-access-wqwjq\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.223886 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-console-oauth-config\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.223962 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-console-config\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.224000 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-oauth-serving-cert\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.224071 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-trusted-ca-bundle\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.224104 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-console-serving-cert\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.225070 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-service-ca\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.226167 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-console-config\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.226874 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-trusted-ca-bundle\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.227827 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-console-serving-cert\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.231078 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-console-oauth-config\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.231141 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-oauth-serving-cert\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.258087 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqwjq\" (UniqueName: \"kubernetes.io/projected/06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723-kube-api-access-wqwjq\") pod \"console-595b748fd5-4v476\" (UID: \"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723\") " pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.311316 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.531382 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f0da1b73-3249-449b-86f9-12206ce93aab-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.532298 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f0da1b73-3249-449b-86f9-12206ce93aab-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:19 crc kubenswrapper[4954]: I1209 17:25:19.798072 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.423961 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"28718e77-f5b5-4d67-afbc-c1157354bc47","Type":"ContainerStarted","Data":"c23918ecab3c55b33ef7105f7a0a5eba7386421da8b3660f44ecb96d397fb104"} Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.585221 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.587072 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.589777 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-5tlgj" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.590214 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.590787 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.591134 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.602267 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.604212 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.655123 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z77m\" (UniqueName: \"kubernetes.io/projected/060bba41-7ec2-4056-9d55-7b9d4ea22b96-kube-api-access-7z77m\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.655190 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/060bba41-7ec2-4056-9d55-7b9d4ea22b96-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.655242 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/060bba41-7ec2-4056-9d55-7b9d4ea22b96-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.655301 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/060bba41-7ec2-4056-9d55-7b9d4ea22b96-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.655739 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.655938 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/060bba41-7ec2-4056-9d55-7b9d4ea22b96-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.656006 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/060bba41-7ec2-4056-9d55-7b9d4ea22b96-config\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.656146 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/060bba41-7ec2-4056-9d55-7b9d4ea22b96-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.757430 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/060bba41-7ec2-4056-9d55-7b9d4ea22b96-config\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.757491 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/060bba41-7ec2-4056-9d55-7b9d4ea22b96-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.757534 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z77m\" (UniqueName: \"kubernetes.io/projected/060bba41-7ec2-4056-9d55-7b9d4ea22b96-kube-api-access-7z77m\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.757564 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/060bba41-7ec2-4056-9d55-7b9d4ea22b96-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.757630 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/060bba41-7ec2-4056-9d55-7b9d4ea22b96-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.757669 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/060bba41-7ec2-4056-9d55-7b9d4ea22b96-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.757721 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.757763 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/060bba41-7ec2-4056-9d55-7b9d4ea22b96-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.758282 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/060bba41-7ec2-4056-9d55-7b9d4ea22b96-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.758341 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.758723 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/060bba41-7ec2-4056-9d55-7b9d4ea22b96-config\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.758930 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/060bba41-7ec2-4056-9d55-7b9d4ea22b96-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.764849 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/060bba41-7ec2-4056-9d55-7b9d4ea22b96-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.767793 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/060bba41-7ec2-4056-9d55-7b9d4ea22b96-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.768095 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/060bba41-7ec2-4056-9d55-7b9d4ea22b96-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.787382 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.797008 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z77m\" (UniqueName: \"kubernetes.io/projected/060bba41-7ec2-4056-9d55-7b9d4ea22b96-kube-api-access-7z77m\") pod \"ovsdbserver-nb-0\" (UID: \"060bba41-7ec2-4056-9d55-7b9d4ea22b96\") " pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:20 crc kubenswrapper[4954]: I1209 17:25:20.920208 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.550949 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-l6s5g"] Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.552655 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.556044 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-7l25r" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.556262 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.556285 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.565376 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l6s5g"] Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.572210 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-ggbdh"] Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.576540 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a5a72d-c7dd-462f-969f-812555bc9ba3-combined-ca-bundle\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.576621 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/94a5a72d-c7dd-462f-969f-812555bc9ba3-var-run-ovn\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.576707 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/94a5a72d-c7dd-462f-969f-812555bc9ba3-ovn-controller-tls-certs\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.576729 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/94a5a72d-c7dd-462f-969f-812555bc9ba3-var-log-ovn\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.576760 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94a5a72d-c7dd-462f-969f-812555bc9ba3-scripts\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.576826 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/94a5a72d-c7dd-462f-969f-812555bc9ba3-var-run\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.576854 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg7kd\" (UniqueName: \"kubernetes.io/projected/94a5a72d-c7dd-462f-969f-812555bc9ba3-kube-api-access-mg7kd\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.588249 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.614645 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-ggbdh"] Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681028 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-var-run\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681121 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/94a5a72d-c7dd-462f-969f-812555bc9ba3-ovn-controller-tls-certs\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681158 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/94a5a72d-c7dd-462f-969f-812555bc9ba3-var-log-ovn\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681194 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-var-log\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681215 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-var-lib\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681245 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94a5a72d-c7dd-462f-969f-812555bc9ba3-scripts\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681267 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfzsh\" (UniqueName: \"kubernetes.io/projected/2352a5dd-0445-4b59-9e8e-4607d055e505-kube-api-access-cfzsh\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681328 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2352a5dd-0445-4b59-9e8e-4607d055e505-scripts\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681391 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/94a5a72d-c7dd-462f-969f-812555bc9ba3-var-run\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681427 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg7kd\" (UniqueName: \"kubernetes.io/projected/94a5a72d-c7dd-462f-969f-812555bc9ba3-kube-api-access-mg7kd\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681458 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a5a72d-c7dd-462f-969f-812555bc9ba3-combined-ca-bundle\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681498 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/94a5a72d-c7dd-462f-969f-812555bc9ba3-var-run-ovn\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.681521 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-etc-ovs\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.682489 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/94a5a72d-c7dd-462f-969f-812555bc9ba3-var-run\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.682512 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/94a5a72d-c7dd-462f-969f-812555bc9ba3-var-log-ovn\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.682608 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/94a5a72d-c7dd-462f-969f-812555bc9ba3-var-run-ovn\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.684389 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94a5a72d-c7dd-462f-969f-812555bc9ba3-scripts\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.686674 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a5a72d-c7dd-462f-969f-812555bc9ba3-combined-ca-bundle\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.690857 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/94a5a72d-c7dd-462f-969f-812555bc9ba3-ovn-controller-tls-certs\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.702637 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg7kd\" (UniqueName: \"kubernetes.io/projected/94a5a72d-c7dd-462f-969f-812555bc9ba3-kube-api-access-mg7kd\") pod \"ovn-controller-l6s5g\" (UID: \"94a5a72d-c7dd-462f-969f-812555bc9ba3\") " pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.783388 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-var-log\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.783440 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-var-lib\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.783467 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfzsh\" (UniqueName: \"kubernetes.io/projected/2352a5dd-0445-4b59-9e8e-4607d055e505-kube-api-access-cfzsh\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.783509 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2352a5dd-0445-4b59-9e8e-4607d055e505-scripts\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.783576 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-etc-ovs\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.783642 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-var-run\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.783770 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-var-run\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.783914 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-var-lib\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.784073 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-etc-ovs\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.784123 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2352a5dd-0445-4b59-9e8e-4607d055e505-var-log\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.805615 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfzsh\" (UniqueName: \"kubernetes.io/projected/2352a5dd-0445-4b59-9e8e-4607d055e505-kube-api-access-cfzsh\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:21 crc kubenswrapper[4954]: I1209 17:25:21.880276 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:22 crc kubenswrapper[4954]: I1209 17:25:22.399063 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2352a5dd-0445-4b59-9e8e-4607d055e505-scripts\") pod \"ovn-controller-ovs-ggbdh\" (UID: \"2352a5dd-0445-4b59-9e8e-4607d055e505\") " pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:22 crc kubenswrapper[4954]: I1209 17:25:22.510160 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.805097 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.807189 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.811334 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.811639 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.811814 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-wx7l5" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.812459 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.834880 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/619c2101-1cbd-4584-a5fd-b72f9963b9d6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.834927 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klj55\" (UniqueName: \"kubernetes.io/projected/619c2101-1cbd-4584-a5fd-b72f9963b9d6-kube-api-access-klj55\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.835033 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/619c2101-1cbd-4584-a5fd-b72f9963b9d6-config\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.835051 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/619c2101-1cbd-4584-a5fd-b72f9963b9d6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.835072 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/619c2101-1cbd-4584-a5fd-b72f9963b9d6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.835155 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/619c2101-1cbd-4584-a5fd-b72f9963b9d6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.835181 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.835203 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/619c2101-1cbd-4584-a5fd-b72f9963b9d6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.836164 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.941157 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/619c2101-1cbd-4584-a5fd-b72f9963b9d6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.951422 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.951557 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/619c2101-1cbd-4584-a5fd-b72f9963b9d6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.951664 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.952228 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/619c2101-1cbd-4584-a5fd-b72f9963b9d6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.952614 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/619c2101-1cbd-4584-a5fd-b72f9963b9d6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.952673 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klj55\" (UniqueName: \"kubernetes.io/projected/619c2101-1cbd-4584-a5fd-b72f9963b9d6-kube-api-access-klj55\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.953017 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/619c2101-1cbd-4584-a5fd-b72f9963b9d6-config\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.953070 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/619c2101-1cbd-4584-a5fd-b72f9963b9d6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.953115 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/619c2101-1cbd-4584-a5fd-b72f9963b9d6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.954581 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/619c2101-1cbd-4584-a5fd-b72f9963b9d6-config\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.955895 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/619c2101-1cbd-4584-a5fd-b72f9963b9d6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.962576 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/619c2101-1cbd-4584-a5fd-b72f9963b9d6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.966887 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/619c2101-1cbd-4584-a5fd-b72f9963b9d6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.969225 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/619c2101-1cbd-4584-a5fd-b72f9963b9d6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.982215 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klj55\" (UniqueName: \"kubernetes.io/projected/619c2101-1cbd-4584-a5fd-b72f9963b9d6-kube-api-access-klj55\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:23 crc kubenswrapper[4954]: I1209 17:25:23.991621 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"619c2101-1cbd-4584-a5fd-b72f9963b9d6\") " pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:24 crc kubenswrapper[4954]: I1209 17:25:24.132146 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:25:24 crc kubenswrapper[4954]: E1209 17:25:24.132957 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:25:24 crc kubenswrapper[4954]: I1209 17:25:24.146547 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:27 crc kubenswrapper[4954]: I1209 17:25:27.040495 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 17:25:33 crc kubenswrapper[4954]: I1209 17:25:33.337112 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 17:25:33 crc kubenswrapper[4954]: I1209 17:25:33.580738 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"72a212eb-2f11-46aa-88ff-6b59b86c5b6c","Type":"ContainerStarted","Data":"cf857834225d1af96ea55cf4172a4be1ac63e1d856ef593388b4a899d1020081"} Dec 09 17:25:33 crc kubenswrapper[4954]: W1209 17:25:33.730288 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0da1b73_3249_449b_86f9_12206ce93aab.slice/crio-53ec0de08d9d6b12288f1fce35364fa6c66400e75e8c152bf38b44f97a107ca5 WatchSource:0}: Error finding container 53ec0de08d9d6b12288f1fce35364fa6c66400e75e8c152bf38b44f97a107ca5: Status 404 returned error can't find the container with id 53ec0de08d9d6b12288f1fce35364fa6c66400e75e8c152bf38b44f97a107ca5 Dec 09 17:25:33 crc kubenswrapper[4954]: E1209 17:25:33.744784 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 09 17:25:33 crc kubenswrapper[4954]: E1209 17:25:33.745182 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z4bkf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-tcbm7_openstack(a142cf6b-3845-4f0a-8e7d-0359af36b21a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:25:33 crc kubenswrapper[4954]: E1209 17:25:33.746580 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" podUID="a142cf6b-3845-4f0a-8e7d-0359af36b21a" Dec 09 17:25:33 crc kubenswrapper[4954]: E1209 17:25:33.812819 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 09 17:25:33 crc kubenswrapper[4954]: E1209 17:25:33.813078 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-59d9w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-kr46n_openstack(086f5371-c7ea-49ba-bd08-2096c1e71d2c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:25:33 crc kubenswrapper[4954]: E1209 17:25:33.815201 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" podUID="086f5371-c7ea-49ba-bd08-2096c1e71d2c" Dec 09 17:25:34 crc kubenswrapper[4954]: I1209 17:25:34.635657 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f0da1b73-3249-449b-86f9-12206ce93aab","Type":"ContainerStarted","Data":"53ec0de08d9d6b12288f1fce35364fa6c66400e75e8c152bf38b44f97a107ca5"} Dec 09 17:25:34 crc kubenswrapper[4954]: I1209 17:25:34.654729 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l6s5g"] Dec 09 17:25:34 crc kubenswrapper[4954]: I1209 17:25:34.847353 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 17:25:34 crc kubenswrapper[4954]: I1209 17:25:34.892473 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 09 17:25:34 crc kubenswrapper[4954]: I1209 17:25:34.907371 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-595b748fd5-4v476"] Dec 09 17:25:34 crc kubenswrapper[4954]: I1209 17:25:34.991491 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.120891 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:25:35 crc kubenswrapper[4954]: E1209 17:25:35.121179 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:25:35 crc kubenswrapper[4954]: W1209 17:25:35.267720 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94a5a72d_c7dd_462f_969f_812555bc9ba3.slice/crio-0319f8b78ee7e80936b35f2bdd84f87a47033e482836f3c44b688f658b9a0aad WatchSource:0}: Error finding container 0319f8b78ee7e80936b35f2bdd84f87a47033e482836f3c44b688f658b9a0aad: Status 404 returned error can't find the container with id 0319f8b78ee7e80936b35f2bdd84f87a47033e482836f3c44b688f658b9a0aad Dec 09 17:25:35 crc kubenswrapper[4954]: W1209 17:25:35.270225 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd42e4546_9029_4930_ab04_7adaab383d5a.slice/crio-be138de92ee9b146fa4626315666580e7502cdee3d30ea36a7189eb785d49b8d WatchSource:0}: Error finding container be138de92ee9b146fa4626315666580e7502cdee3d30ea36a7189eb785d49b8d: Status 404 returned error can't find the container with id be138de92ee9b146fa4626315666580e7502cdee3d30ea36a7189eb785d49b8d Dec 09 17:25:35 crc kubenswrapper[4954]: W1209 17:25:35.280748 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06abb4cf_5fcd_4ecc_b5e3_fcd9c84fe723.slice/crio-0f46ea1d6ab42aeb76086019e0e50ec830b10ee0d59dce724ad1c0b8ee3cb835 WatchSource:0}: Error finding container 0f46ea1d6ab42aeb76086019e0e50ec830b10ee0d59dce724ad1c0b8ee3cb835: Status 404 returned error can't find the container with id 0f46ea1d6ab42aeb76086019e0e50ec830b10ee0d59dce724ad1c0b8ee3cb835 Dec 09 17:25:35 crc kubenswrapper[4954]: W1209 17:25:35.285279 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod060bba41_7ec2_4056_9d55_7b9d4ea22b96.slice/crio-d16549aaae5e14e09ec94e3cf51128552aec28dabf42cfc3deaacaa8c8cb87a9 WatchSource:0}: Error finding container d16549aaae5e14e09ec94e3cf51128552aec28dabf42cfc3deaacaa8c8cb87a9: Status 404 returned error can't find the container with id d16549aaae5e14e09ec94e3cf51128552aec28dabf42cfc3deaacaa8c8cb87a9 Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.353775 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.365964 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.377933 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.402412 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5"] Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.450263 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-dns-svc\") pod \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.450923 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4bkf\" (UniqueName: \"kubernetes.io/projected/a142cf6b-3845-4f0a-8e7d-0359af36b21a-kube-api-access-z4bkf\") pod \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.451017 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a142cf6b-3845-4f0a-8e7d-0359af36b21a" (UID: "a142cf6b-3845-4f0a-8e7d-0359af36b21a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.451066 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-config\") pod \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\" (UID: \"a142cf6b-3845-4f0a-8e7d-0359af36b21a\") " Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.451722 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.452106 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-config" (OuterVolumeSpecName: "config") pod "a142cf6b-3845-4f0a-8e7d-0359af36b21a" (UID: "a142cf6b-3845-4f0a-8e7d-0359af36b21a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.458585 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a142cf6b-3845-4f0a-8e7d-0359af36b21a-kube-api-access-z4bkf" (OuterVolumeSpecName: "kube-api-access-z4bkf") pod "a142cf6b-3845-4f0a-8e7d-0359af36b21a" (UID: "a142cf6b-3845-4f0a-8e7d-0359af36b21a"). InnerVolumeSpecName "kube-api-access-z4bkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.555208 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4bkf\" (UniqueName: \"kubernetes.io/projected/a142cf6b-3845-4f0a-8e7d-0359af36b21a-kube-api-access-z4bkf\") on node \"crc\" DevicePath \"\"" Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.555252 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a142cf6b-3845-4f0a-8e7d-0359af36b21a-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.658962 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d42e4546-9029-4930-ab04-7adaab383d5a","Type":"ContainerStarted","Data":"be138de92ee9b146fa4626315666580e7502cdee3d30ea36a7189eb785d49b8d"} Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.661945 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l6s5g" event={"ID":"94a5a72d-c7dd-462f-969f-812555bc9ba3","Type":"ContainerStarted","Data":"0319f8b78ee7e80936b35f2bdd84f87a47033e482836f3c44b688f658b9a0aad"} Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.663844 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"bb8de77c-60eb-4e0d-abdf-d6bc9f111031","Type":"ContainerStarted","Data":"eb257d2ad72dd9d588053305e0e593e44b80fb1e30df8733a0be054608505d0d"} Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.665878 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-595b748fd5-4v476" event={"ID":"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723","Type":"ContainerStarted","Data":"0f46ea1d6ab42aeb76086019e0e50ec830b10ee0d59dce724ad1c0b8ee3cb835"} Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.668448 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3645ed14-3d97-496e-a683-2e56ea33fec9","Type":"ContainerStarted","Data":"f09af2ede48c6c2cac1998e75e7a105a2e32d41d51df5ef6f04a13d9226440c0"} Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.670207 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" event={"ID":"a142cf6b-3845-4f0a-8e7d-0359af36b21a","Type":"ContainerDied","Data":"92cf589e1bf97a0db61ed2fc039a6483b25af8c9fb9093dba80bc7658338636b"} Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.670257 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tcbm7" Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.673488 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"060bba41-7ec2-4056-9d55-7b9d4ea22b96","Type":"ContainerStarted","Data":"d16549aaae5e14e09ec94e3cf51128552aec28dabf42cfc3deaacaa8c8cb87a9"} Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.676034 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"28718e77-f5b5-4d67-afbc-c1157354bc47","Type":"ContainerStarted","Data":"fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad"} Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.679096 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"197b01d4-b9ac-4dc6-ac33-4638ece0241f","Type":"ContainerStarted","Data":"bf2273bd549cbaf680c0f4b2c48a6fc8575f1f32da19279d144f3f5adfefcb83"} Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.680159 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" event={"ID":"300e9948-95ee-44e4-b816-16664d86e2e2","Type":"ContainerStarted","Data":"ff2aecaffaff669ccddf2f51f742a8907c0724b4934a0ad9527d164bdfedf341"} Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.762510 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tcbm7"] Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.770835 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tcbm7"] Dec 09 17:25:35 crc kubenswrapper[4954]: I1209 17:25:35.844155 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-ggbdh"] Dec 09 17:25:36 crc kubenswrapper[4954]: I1209 17:25:36.106037 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 17:25:36 crc kubenswrapper[4954]: I1209 17:25:36.157709 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a142cf6b-3845-4f0a-8e7d-0359af36b21a" path="/var/lib/kubelet/pods/a142cf6b-3845-4f0a-8e7d-0359af36b21a/volumes" Dec 09 17:25:36 crc kubenswrapper[4954]: W1209 17:25:36.263636 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod619c2101_1cbd_4584_a5fd_b72f9963b9d6.slice/crio-5983327efa4085696b453f7b9e877a98cf81b70d25fa152dbef1a01864372f13 WatchSource:0}: Error finding container 5983327efa4085696b453f7b9e877a98cf81b70d25fa152dbef1a01864372f13: Status 404 returned error can't find the container with id 5983327efa4085696b453f7b9e877a98cf81b70d25fa152dbef1a01864372f13 Dec 09 17:25:36 crc kubenswrapper[4954]: I1209 17:25:36.701298 4954 generic.go:334] "Generic (PLEG): container finished" podID="62728408-1835-4e60-9d36-dac31093b2b9" containerID="e05cd1b43a9c0079a779bac0e7d0a6b65db1dd596377a17c5c40a25bebbfdb2a" exitCode=0 Dec 09 17:25:36 crc kubenswrapper[4954]: I1209 17:25:36.702108 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" event={"ID":"62728408-1835-4e60-9d36-dac31093b2b9","Type":"ContainerDied","Data":"e05cd1b43a9c0079a779bac0e7d0a6b65db1dd596377a17c5c40a25bebbfdb2a"} Dec 09 17:25:36 crc kubenswrapper[4954]: I1209 17:25:36.704962 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ggbdh" event={"ID":"2352a5dd-0445-4b59-9e8e-4607d055e505","Type":"ContainerStarted","Data":"af95f0e63e8b85b94e879f050a640e65fbd3bbca054032a6238943559d6911d2"} Dec 09 17:25:36 crc kubenswrapper[4954]: I1209 17:25:36.706210 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"619c2101-1cbd-4584-a5fd-b72f9963b9d6","Type":"ContainerStarted","Data":"5983327efa4085696b453f7b9e877a98cf81b70d25fa152dbef1a01864372f13"} Dec 09 17:25:36 crc kubenswrapper[4954]: I1209 17:25:36.717518 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-595b748fd5-4v476" event={"ID":"06abb4cf-5fcd-4ecc-b5e3-fcd9c84fe723","Type":"ContainerStarted","Data":"bdbaa9532435473968e1873f35253f7c38e4ab16bc3479715cf636e01f58e824"} Dec 09 17:25:36 crc kubenswrapper[4954]: I1209 17:25:36.762213 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-595b748fd5-4v476" podStartSLOduration=18.762182487 podStartE2EDuration="18.762182487s" podCreationTimestamp="2025-12-09 17:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:25:36.756562891 +0000 UTC m=+1733.144736901" watchObservedRunningTime="2025-12-09 17:25:36.762182487 +0000 UTC m=+1733.150356317" Dec 09 17:25:39 crc kubenswrapper[4954]: I1209 17:25:39.311823 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:39 crc kubenswrapper[4954]: I1209 17:25:39.312326 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:39 crc kubenswrapper[4954]: I1209 17:25:39.320017 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:39 crc kubenswrapper[4954]: I1209 17:25:39.760573 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-595b748fd5-4v476" Dec 09 17:25:39 crc kubenswrapper[4954]: I1209 17:25:39.841801 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-bfd664845-r74kf"] Dec 09 17:25:46 crc kubenswrapper[4954]: I1209 17:25:46.120993 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:25:46 crc kubenswrapper[4954]: E1209 17:25:46.122202 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.486864 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.874814 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d42e4546-9029-4930-ab04-7adaab383d5a","Type":"ContainerStarted","Data":"8f6763ca1554588cb7faca2edc833a1ed2286e0a091ec23b4d21b9133ca13d86"} Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.877378 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l6s5g" event={"ID":"94a5a72d-c7dd-462f-969f-812555bc9ba3","Type":"ContainerStarted","Data":"f451c0a7c941af04a2a3171faa887cd93f2edb59e5a5dee1cceef10e350ee848"} Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.877537 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-l6s5g" Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.878950 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"72a212eb-2f11-46aa-88ff-6b59b86c5b6c","Type":"ContainerStarted","Data":"bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58"} Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.879963 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.882336 4954 generic.go:334] "Generic (PLEG): container finished" podID="086f5371-c7ea-49ba-bd08-2096c1e71d2c" containerID="08b8b7079738285e018a76e393a3ae16abec97a138723bbcaada12a1b500fb5b" exitCode=0 Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.882403 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" event={"ID":"086f5371-c7ea-49ba-bd08-2096c1e71d2c","Type":"ContainerDied","Data":"08b8b7079738285e018a76e393a3ae16abec97a138723bbcaada12a1b500fb5b"} Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.885903 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" event={"ID":"62728408-1835-4e60-9d36-dac31093b2b9","Type":"ContainerStarted","Data":"468ee6241122089955dc94786927719be576853d09c9219071a90a54f2dde898"} Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.886027 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.913070 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" event={"ID":"300e9948-95ee-44e4-b816-16664d86e2e2","Type":"ContainerStarted","Data":"d3f39941a2eae3266cc7b03102e44f0393f61e9a11ed8d56f31bb08e888081fd"} Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.927256 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ggbdh" event={"ID":"2352a5dd-0445-4b59-9e8e-4607d055e505","Type":"ContainerStarted","Data":"94e176e36ef66844e01ca668544f3daafe152cc25d62022b46cc190b61a95f68"} Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.944861 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"060bba41-7ec2-4056-9d55-7b9d4ea22b96","Type":"ContainerStarted","Data":"89e75c23d6d6d1f98e97c42967282c2e1c5b1f7a51fc217a89f29d89f97a8228"} Dec 09 17:25:50 crc kubenswrapper[4954]: I1209 17:25:50.970052 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mmts5" podStartSLOduration=19.557579477 podStartE2EDuration="32.970022849s" podCreationTimestamp="2025-12-09 17:25:18 +0000 UTC" firstStartedPulling="2025-12-09 17:25:35.614529599 +0000 UTC m=+1732.002703419" lastFinishedPulling="2025-12-09 17:25:49.026972971 +0000 UTC m=+1745.415146791" observedRunningTime="2025-12-09 17:25:50.956820215 +0000 UTC m=+1747.344994025" watchObservedRunningTime="2025-12-09 17:25:50.970022849 +0000 UTC m=+1747.358196669" Dec 09 17:25:51 crc kubenswrapper[4954]: I1209 17:25:51.023511 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" podStartSLOduration=18.818329829 podStartE2EDuration="41.023490992s" podCreationTimestamp="2025-12-09 17:25:10 +0000 UTC" firstStartedPulling="2025-12-09 17:25:11.722872509 +0000 UTC m=+1708.111046329" lastFinishedPulling="2025-12-09 17:25:33.928033672 +0000 UTC m=+1730.316207492" observedRunningTime="2025-12-09 17:25:50.991173001 +0000 UTC m=+1747.379346821" watchObservedRunningTime="2025-12-09 17:25:51.023490992 +0000 UTC m=+1747.411664812" Dec 09 17:25:51 crc kubenswrapper[4954]: I1209 17:25:51.025377 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-l6s5g" podStartSLOduration=16.208663713 podStartE2EDuration="30.025370692s" podCreationTimestamp="2025-12-09 17:25:21 +0000 UTC" firstStartedPulling="2025-12-09 17:25:35.270668223 +0000 UTC m=+1731.658842043" lastFinishedPulling="2025-12-09 17:25:49.087375182 +0000 UTC m=+1745.475549022" observedRunningTime="2025-12-09 17:25:51.018995942 +0000 UTC m=+1747.407169762" watchObservedRunningTime="2025-12-09 17:25:51.025370692 +0000 UTC m=+1747.413544512" Dec 09 17:25:51 crc kubenswrapper[4954]: I1209 17:25:51.086138 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=17.902210289 podStartE2EDuration="34.086108163s" podCreationTimestamp="2025-12-09 17:25:17 +0000 UTC" firstStartedPulling="2025-12-09 17:25:32.858060656 +0000 UTC m=+1729.246234476" lastFinishedPulling="2025-12-09 17:25:49.04195854 +0000 UTC m=+1745.430132350" observedRunningTime="2025-12-09 17:25:51.047311959 +0000 UTC m=+1747.435485779" watchObservedRunningTime="2025-12-09 17:25:51.086108163 +0000 UTC m=+1747.474281973" Dec 09 17:25:51 crc kubenswrapper[4954]: E1209 17:25:51.285725 4954 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod086f5371_c7ea_49ba_bd08_2096c1e71d2c.slice/crio-08b8b7079738285e018a76e393a3ae16abec97a138723bbcaada12a1b500fb5b.scope\": RecentStats: unable to find data in memory cache]" Dec 09 17:25:51 crc kubenswrapper[4954]: I1209 17:25:51.957038 4954 generic.go:334] "Generic (PLEG): container finished" podID="2352a5dd-0445-4b59-9e8e-4607d055e505" containerID="94e176e36ef66844e01ca668544f3daafe152cc25d62022b46cc190b61a95f68" exitCode=0 Dec 09 17:25:51 crc kubenswrapper[4954]: I1209 17:25:51.957235 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ggbdh" event={"ID":"2352a5dd-0445-4b59-9e8e-4607d055e505","Type":"ContainerDied","Data":"94e176e36ef66844e01ca668544f3daafe152cc25d62022b46cc190b61a95f68"} Dec 09 17:25:51 crc kubenswrapper[4954]: I1209 17:25:51.966692 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"bb8de77c-60eb-4e0d-abdf-d6bc9f111031","Type":"ContainerStarted","Data":"55322df8525b9f2af14402bd28296ea25d107e6a3623d153fc261f755135c464"} Dec 09 17:25:51 crc kubenswrapper[4954]: I1209 17:25:51.966760 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 09 17:25:51 crc kubenswrapper[4954]: I1209 17:25:51.972154 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"619c2101-1cbd-4584-a5fd-b72f9963b9d6","Type":"ContainerStarted","Data":"a0cded4b480199c419bd09f438a3c5c5d5a99cc1392e4a2df8392a07797d8246"} Dec 09 17:25:51 crc kubenswrapper[4954]: I1209 17:25:51.980536 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3645ed14-3d97-496e-a683-2e56ea33fec9","Type":"ContainerStarted","Data":"e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6"} Dec 09 17:25:51 crc kubenswrapper[4954]: I1209 17:25:51.995187 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"197b01d4-b9ac-4dc6-ac33-4638ece0241f","Type":"ContainerStarted","Data":"586710a3461c0039498b5b7de5c4fd5023c2558a1f4997e1f67669ccddb6c932"} Dec 09 17:25:52 crc kubenswrapper[4954]: I1209 17:25:52.033696 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=24.122705758 podStartE2EDuration="38.033671327s" podCreationTimestamp="2025-12-09 17:25:14 +0000 UTC" firstStartedPulling="2025-12-09 17:25:35.276797785 +0000 UTC m=+1731.664971605" lastFinishedPulling="2025-12-09 17:25:49.187763354 +0000 UTC m=+1745.575937174" observedRunningTime="2025-12-09 17:25:52.013667191 +0000 UTC m=+1748.401841011" watchObservedRunningTime="2025-12-09 17:25:52.033671327 +0000 UTC m=+1748.421845147" Dec 09 17:25:53 crc kubenswrapper[4954]: I1209 17:25:53.009018 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f0da1b73-3249-449b-86f9-12206ce93aab","Type":"ContainerStarted","Data":"bfa38e1cf3605a9d66bcda5a2cb2ed06a8bfbea9eb7f3287ed2e14d8d3509a38"} Dec 09 17:25:53 crc kubenswrapper[4954]: I1209 17:25:53.017094 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" event={"ID":"086f5371-c7ea-49ba-bd08-2096c1e71d2c","Type":"ContainerStarted","Data":"6e1f41f878eced48b084afec2fcdb5a0cc6c81e21a0b138ad2ed41fb3990acb6"} Dec 09 17:25:53 crc kubenswrapper[4954]: I1209 17:25:53.017653 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:53 crc kubenswrapper[4954]: I1209 17:25:53.020515 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ggbdh" event={"ID":"2352a5dd-0445-4b59-9e8e-4607d055e505","Type":"ContainerStarted","Data":"8cbde35ba8c3e0a3df5ef8e01f9f0d3f0e35bc7ca46470bf09245b4afd2ee87e"} Dec 09 17:25:53 crc kubenswrapper[4954]: I1209 17:25:53.060891 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" podStartSLOduration=-9223371993.79391 podStartE2EDuration="43.060867614s" podCreationTimestamp="2025-12-09 17:25:10 +0000 UTC" firstStartedPulling="2025-12-09 17:25:11.394679915 +0000 UTC m=+1707.782853735" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:25:53.054221576 +0000 UTC m=+1749.442395406" watchObservedRunningTime="2025-12-09 17:25:53.060867614 +0000 UTC m=+1749.449041434" Dec 09 17:25:53 crc kubenswrapper[4954]: I1209 17:25:53.944582 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-g9rnm"] Dec 09 17:25:53 crc kubenswrapper[4954]: I1209 17:25:53.947614 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:53 crc kubenswrapper[4954]: I1209 17:25:53.960209 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.031455 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-g9rnm"] Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.047828 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/27f44545-77d2-4048-9bfc-cd8504a1b837-ovn-rundir\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.051964 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj7hz\" (UniqueName: \"kubernetes.io/projected/27f44545-77d2-4048-9bfc-cd8504a1b837-kube-api-access-wj7hz\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.052070 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f44545-77d2-4048-9bfc-cd8504a1b837-combined-ca-bundle\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.052276 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/27f44545-77d2-4048-9bfc-cd8504a1b837-ovs-rundir\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.052738 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/27f44545-77d2-4048-9bfc-cd8504a1b837-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.052854 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27f44545-77d2-4048-9bfc-cd8504a1b837-config\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.159174 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/27f44545-77d2-4048-9bfc-cd8504a1b837-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.159267 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27f44545-77d2-4048-9bfc-cd8504a1b837-config\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.159371 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/27f44545-77d2-4048-9bfc-cd8504a1b837-ovn-rundir\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.159433 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj7hz\" (UniqueName: \"kubernetes.io/projected/27f44545-77d2-4048-9bfc-cd8504a1b837-kube-api-access-wj7hz\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.159472 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f44545-77d2-4048-9bfc-cd8504a1b837-combined-ca-bundle\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.159495 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/27f44545-77d2-4048-9bfc-cd8504a1b837-ovs-rundir\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.162780 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/27f44545-77d2-4048-9bfc-cd8504a1b837-ovn-rundir\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.164231 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27f44545-77d2-4048-9bfc-cd8504a1b837-config\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.164485 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/27f44545-77d2-4048-9bfc-cd8504a1b837-ovs-rundir\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.172553 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f44545-77d2-4048-9bfc-cd8504a1b837-combined-ca-bundle\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.205145 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-x9rc7"] Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.205455 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" podUID="62728408-1835-4e60-9d36-dac31093b2b9" containerName="dnsmasq-dns" containerID="cri-o://468ee6241122089955dc94786927719be576853d09c9219071a90a54f2dde898" gracePeriod=10 Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.208360 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/27f44545-77d2-4048-9bfc-cd8504a1b837-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.228810 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj7hz\" (UniqueName: \"kubernetes.io/projected/27f44545-77d2-4048-9bfc-cd8504a1b837-kube-api-access-wj7hz\") pod \"ovn-controller-metrics-g9rnm\" (UID: \"27f44545-77d2-4048-9bfc-cd8504a1b837\") " pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.273661 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-599bw"] Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.275631 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.290077 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.304775 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-599bw"] Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.318139 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-g9rnm" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.369769 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.369841 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w92l2\" (UniqueName: \"kubernetes.io/projected/6d66ceb0-8567-43c3-bbda-46693b51be82-kube-api-access-w92l2\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.369908 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.369940 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-config\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.472138 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.472533 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w92l2\" (UniqueName: \"kubernetes.io/projected/6d66ceb0-8567-43c3-bbda-46693b51be82-kube-api-access-w92l2\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.472636 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.472683 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-config\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.473629 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-config\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.474223 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.474730 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.500815 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w92l2\" (UniqueName: \"kubernetes.io/projected/6d66ceb0-8567-43c3-bbda-46693b51be82-kube-api-access-w92l2\") pod \"dnsmasq-dns-7fd796d7df-599bw\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.622197 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.671902 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kr46n"] Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.723203 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-n5hmc"] Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.725065 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.728686 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.741817 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-n5hmc"] Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.779848 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.779949 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.780338 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.780644 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-config\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.780804 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bspsr\" (UniqueName: \"kubernetes.io/projected/a767b973-ed53-4675-93a8-db722496f8d7-kube-api-access-bspsr\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.884077 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-config\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.884163 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bspsr\" (UniqueName: \"kubernetes.io/projected/a767b973-ed53-4675-93a8-db722496f8d7-kube-api-access-bspsr\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.884237 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.884263 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.884328 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.885397 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-config\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.885451 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.885745 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.885785 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:54 crc kubenswrapper[4954]: I1209 17:25:54.917932 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bspsr\" (UniqueName: \"kubernetes.io/projected/a767b973-ed53-4675-93a8-db722496f8d7-kube-api-access-bspsr\") pod \"dnsmasq-dns-86db49b7ff-n5hmc\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.112295 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.116437 4954 generic.go:334] "Generic (PLEG): container finished" podID="62728408-1835-4e60-9d36-dac31093b2b9" containerID="468ee6241122089955dc94786927719be576853d09c9219071a90a54f2dde898" exitCode=0 Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.116512 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" event={"ID":"62728408-1835-4e60-9d36-dac31093b2b9","Type":"ContainerDied","Data":"468ee6241122089955dc94786927719be576853d09c9219071a90a54f2dde898"} Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.118103 4954 generic.go:334] "Generic (PLEG): container finished" podID="d42e4546-9029-4930-ab04-7adaab383d5a" containerID="8f6763ca1554588cb7faca2edc833a1ed2286e0a091ec23b4d21b9133ca13d86" exitCode=0 Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.118772 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d42e4546-9029-4930-ab04-7adaab383d5a","Type":"ContainerDied","Data":"8f6763ca1554588cb7faca2edc833a1ed2286e0a091ec23b4d21b9133ca13d86"} Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.761246 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.841270 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-config\") pod \"62728408-1835-4e60-9d36-dac31093b2b9\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.841826 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-dns-svc\") pod \"62728408-1835-4e60-9d36-dac31093b2b9\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.841955 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gp684\" (UniqueName: \"kubernetes.io/projected/62728408-1835-4e60-9d36-dac31093b2b9-kube-api-access-gp684\") pod \"62728408-1835-4e60-9d36-dac31093b2b9\" (UID: \"62728408-1835-4e60-9d36-dac31093b2b9\") " Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.854514 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62728408-1835-4e60-9d36-dac31093b2b9-kube-api-access-gp684" (OuterVolumeSpecName: "kube-api-access-gp684") pod "62728408-1835-4e60-9d36-dac31093b2b9" (UID: "62728408-1835-4e60-9d36-dac31093b2b9"). InnerVolumeSpecName "kube-api-access-gp684". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:25:55 crc kubenswrapper[4954]: I1209 17:25:55.944546 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gp684\" (UniqueName: \"kubernetes.io/projected/62728408-1835-4e60-9d36-dac31093b2b9-kube-api-access-gp684\") on node \"crc\" DevicePath \"\"" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:55.969459 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "62728408-1835-4e60-9d36-dac31093b2b9" (UID: "62728408-1835-4e60-9d36-dac31093b2b9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:55.994908 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-config" (OuterVolumeSpecName: "config") pod "62728408-1835-4e60-9d36-dac31093b2b9" (UID: "62728408-1835-4e60-9d36-dac31093b2b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.047690 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.047727 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/62728408-1835-4e60-9d36-dac31093b2b9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.070464 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-g9rnm"] Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.140150 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"619c2101-1cbd-4584-a5fd-b72f9963b9d6","Type":"ContainerStarted","Data":"22bb6833cefdfb31699134dc7a028725ba37e25b0f861e4888b3d7b6438fe5bb"} Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.154065 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.157426 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-x9rc7" event={"ID":"62728408-1835-4e60-9d36-dac31093b2b9","Type":"ContainerDied","Data":"d9f741b2bbcaeb18ee1a62357ff7c5061357e5714010c73c58a58c7f3043bbe2"} Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.157509 4954 scope.go:117] "RemoveContainer" containerID="468ee6241122089955dc94786927719be576853d09c9219071a90a54f2dde898" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.170949 4954 generic.go:334] "Generic (PLEG): container finished" podID="197b01d4-b9ac-4dc6-ac33-4638ece0241f" containerID="586710a3461c0039498b5b7de5c4fd5023c2558a1f4997e1f67669ccddb6c932" exitCode=0 Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.171019 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"197b01d4-b9ac-4dc6-ac33-4638ece0241f","Type":"ContainerDied","Data":"586710a3461c0039498b5b7de5c4fd5023c2558a1f4997e1f67669ccddb6c932"} Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.176821 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"060bba41-7ec2-4056-9d55-7b9d4ea22b96","Type":"ContainerStarted","Data":"6a441434a612a8f9020f38ce85be0d4658985c30ba63875e1e146e9eb4d4756a"} Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.182760 4954 scope.go:117] "RemoveContainer" containerID="e05cd1b43a9c0079a779bac0e7d0a6b65db1dd596377a17c5c40a25bebbfdb2a" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.184137 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=14.855253759 podStartE2EDuration="34.184111748s" podCreationTimestamp="2025-12-09 17:25:22 +0000 UTC" firstStartedPulling="2025-12-09 17:25:36.283779059 +0000 UTC m=+1732.671952879" lastFinishedPulling="2025-12-09 17:25:55.612637048 +0000 UTC m=+1752.000810868" observedRunningTime="2025-12-09 17:25:56.16243259 +0000 UTC m=+1752.550606430" watchObservedRunningTime="2025-12-09 17:25:56.184111748 +0000 UTC m=+1752.572285568" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.195939 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ggbdh" event={"ID":"2352a5dd-0445-4b59-9e8e-4607d055e505","Type":"ContainerStarted","Data":"b9511594d47385fa1cd44ff817ce09330fdef2a3344c2035678c9a0fc7866ed1"} Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.197127 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.197165 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.207571 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-x9rc7"] Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.208868 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d42e4546-9029-4930-ab04-7adaab383d5a","Type":"ContainerStarted","Data":"45ee60f893f3a0cc182175f708efe2ac2ae9f138478ff687964daf3f116f5ad0"} Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.208964 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" podUID="086f5371-c7ea-49ba-bd08-2096c1e71d2c" containerName="dnsmasq-dns" containerID="cri-o://6e1f41f878eced48b084afec2fcdb5a0cc6c81e21a0b138ad2ed41fb3990acb6" gracePeriod=10 Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.224304 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-x9rc7"] Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.310628 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=17.023741359 podStartE2EDuration="37.310577118s" podCreationTimestamp="2025-12-09 17:25:19 +0000 UTC" firstStartedPulling="2025-12-09 17:25:35.288564424 +0000 UTC m=+1731.676738244" lastFinishedPulling="2025-12-09 17:25:55.575400183 +0000 UTC m=+1751.963574003" observedRunningTime="2025-12-09 17:25:56.219916449 +0000 UTC m=+1752.608090269" watchObservedRunningTime="2025-12-09 17:25:56.310577118 +0000 UTC m=+1752.698750938" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.342474 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-ggbdh" podStartSLOduration=22.364549506 podStartE2EDuration="35.342448506s" podCreationTimestamp="2025-12-09 17:25:21 +0000 UTC" firstStartedPulling="2025-12-09 17:25:36.138370647 +0000 UTC m=+1732.526544467" lastFinishedPulling="2025-12-09 17:25:49.116269647 +0000 UTC m=+1745.504443467" observedRunningTime="2025-12-09 17:25:56.292934365 +0000 UTC m=+1752.681108185" watchObservedRunningTime="2025-12-09 17:25:56.342448506 +0000 UTC m=+1752.730622326" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.356034 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=30.549816532 podStartE2EDuration="44.354704979s" podCreationTimestamp="2025-12-09 17:25:12 +0000 UTC" firstStartedPulling="2025-12-09 17:25:35.28139031 +0000 UTC m=+1731.669564120" lastFinishedPulling="2025-12-09 17:25:49.086278747 +0000 UTC m=+1745.474452567" observedRunningTime="2025-12-09 17:25:56.322977886 +0000 UTC m=+1752.711151726" watchObservedRunningTime="2025-12-09 17:25:56.354704979 +0000 UTC m=+1752.742878799" Dec 09 17:25:56 crc kubenswrapper[4954]: I1209 17:25:56.921120 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.083385 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.147265 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.191106 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.230630 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"197b01d4-b9ac-4dc6-ac33-4638ece0241f","Type":"ContainerStarted","Data":"fe5f749631e88e12ec082cc1035f9e03a14cef4ed56c8a7bb6c063ac5e4ec107"} Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.233837 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-g9rnm" event={"ID":"27f44545-77d2-4048-9bfc-cd8504a1b837","Type":"ContainerStarted","Data":"a633a8dc2fe960c7904ec8cab59c49e151d69cfcb3c7c11bdecd1aa4c75531be"} Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.233901 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-g9rnm" event={"ID":"27f44545-77d2-4048-9bfc-cd8504a1b837","Type":"ContainerStarted","Data":"a1a2633946c2b78581beaef685bef63f286a23fe64d9809ddc954cf9c9601fa2"} Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.236553 4954 generic.go:334] "Generic (PLEG): container finished" podID="086f5371-c7ea-49ba-bd08-2096c1e71d2c" containerID="6e1f41f878eced48b084afec2fcdb5a0cc6c81e21a0b138ad2ed41fb3990acb6" exitCode=0 Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.237131 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" event={"ID":"086f5371-c7ea-49ba-bd08-2096c1e71d2c","Type":"ContainerDied","Data":"6e1f41f878eced48b084afec2fcdb5a0cc6c81e21a0b138ad2ed41fb3990acb6"} Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.237451 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.237683 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.261156 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=30.699765401 podStartE2EDuration="44.261126265s" podCreationTimestamp="2025-12-09 17:25:13 +0000 UTC" firstStartedPulling="2025-12-09 17:25:35.598099674 +0000 UTC m=+1731.986273494" lastFinishedPulling="2025-12-09 17:25:49.159460528 +0000 UTC m=+1745.547634358" observedRunningTime="2025-12-09 17:25:57.255024574 +0000 UTC m=+1753.643198394" watchObservedRunningTime="2025-12-09 17:25:57.261126265 +0000 UTC m=+1753.649300085" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.287448 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-g9rnm" podStartSLOduration=4.2874156580000005 podStartE2EDuration="4.287415658s" podCreationTimestamp="2025-12-09 17:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:25:57.273946246 +0000 UTC m=+1753.662120076" watchObservedRunningTime="2025-12-09 17:25:57.287415658 +0000 UTC m=+1753.675589478" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.337323 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.376325 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.400989 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-n5hmc"] Dec 09 17:25:57 crc kubenswrapper[4954]: W1209 17:25:57.435637 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d66ceb0_8567_43c3_bbda_46693b51be82.slice/crio-f80e40e8b9f3835184d2c6014ce5b9e19e6d6bab730ed3a5902764a203df0ddb WatchSource:0}: Error finding container f80e40e8b9f3835184d2c6014ce5b9e19e6d6bab730ed3a5902764a203df0ddb: Status 404 returned error can't find the container with id f80e40e8b9f3835184d2c6014ce5b9e19e6d6bab730ed3a5902764a203df0ddb Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.458938 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-599bw"] Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.706729 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 09 17:25:57 crc kubenswrapper[4954]: E1209 17:25:57.714172 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62728408-1835-4e60-9d36-dac31093b2b9" containerName="dnsmasq-dns" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.714205 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="62728408-1835-4e60-9d36-dac31093b2b9" containerName="dnsmasq-dns" Dec 09 17:25:57 crc kubenswrapper[4954]: E1209 17:25:57.714231 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62728408-1835-4e60-9d36-dac31093b2b9" containerName="init" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.714238 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="62728408-1835-4e60-9d36-dac31093b2b9" containerName="init" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.714425 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="62728408-1835-4e60-9d36-dac31093b2b9" containerName="dnsmasq-dns" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.715618 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.719009 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.719513 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-d9drr" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.719690 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.719831 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.740085 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.755503 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.841174 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.856990 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-config\") pod \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.857323 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-dns-svc\") pod \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.857437 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59d9w\" (UniqueName: \"kubernetes.io/projected/086f5371-c7ea-49ba-bd08-2096c1e71d2c-kube-api-access-59d9w\") pod \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\" (UID: \"086f5371-c7ea-49ba-bd08-2096c1e71d2c\") " Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.857778 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b77bf96d-5422-416f-ba9d-cd20e94d991e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.857832 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b77bf96d-5422-416f-ba9d-cd20e94d991e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.857888 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b77bf96d-5422-416f-ba9d-cd20e94d991e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.857936 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b77bf96d-5422-416f-ba9d-cd20e94d991e-scripts\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.857960 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b77bf96d-5422-416f-ba9d-cd20e94d991e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.858018 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5z49\" (UniqueName: \"kubernetes.io/projected/b77bf96d-5422-416f-ba9d-cd20e94d991e-kube-api-access-c5z49\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.858133 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77bf96d-5422-416f-ba9d-cd20e94d991e-config\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.883931 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/086f5371-c7ea-49ba-bd08-2096c1e71d2c-kube-api-access-59d9w" (OuterVolumeSpecName: "kube-api-access-59d9w") pod "086f5371-c7ea-49ba-bd08-2096c1e71d2c" (UID: "086f5371-c7ea-49ba-bd08-2096c1e71d2c"). InnerVolumeSpecName "kube-api-access-59d9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.962852 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b77bf96d-5422-416f-ba9d-cd20e94d991e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.963292 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b77bf96d-5422-416f-ba9d-cd20e94d991e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.963359 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b77bf96d-5422-416f-ba9d-cd20e94d991e-scripts\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.963380 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b77bf96d-5422-416f-ba9d-cd20e94d991e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.963471 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5z49\" (UniqueName: \"kubernetes.io/projected/b77bf96d-5422-416f-ba9d-cd20e94d991e-kube-api-access-c5z49\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.963651 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77bf96d-5422-416f-ba9d-cd20e94d991e-config\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.963718 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b77bf96d-5422-416f-ba9d-cd20e94d991e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.963801 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59d9w\" (UniqueName: \"kubernetes.io/projected/086f5371-c7ea-49ba-bd08-2096c1e71d2c-kube-api-access-59d9w\") on node \"crc\" DevicePath \"\"" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.964035 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b77bf96d-5422-416f-ba9d-cd20e94d991e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.965289 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b77bf96d-5422-416f-ba9d-cd20e94d991e-scripts\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.965682 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77bf96d-5422-416f-ba9d-cd20e94d991e-config\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.970191 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b77bf96d-5422-416f-ba9d-cd20e94d991e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.971891 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b77bf96d-5422-416f-ba9d-cd20e94d991e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.992943 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b77bf96d-5422-416f-ba9d-cd20e94d991e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:57 crc kubenswrapper[4954]: I1209 17:25:57.994429 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5z49\" (UniqueName: \"kubernetes.io/projected/b77bf96d-5422-416f-ba9d-cd20e94d991e-kube-api-access-c5z49\") pod \"ovn-northd-0\" (UID: \"b77bf96d-5422-416f-ba9d-cd20e94d991e\") " pod="openstack/ovn-northd-0" Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.029854 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-config" (OuterVolumeSpecName: "config") pod "086f5371-c7ea-49ba-bd08-2096c1e71d2c" (UID: "086f5371-c7ea-49ba-bd08-2096c1e71d2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.042948 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "086f5371-c7ea-49ba-bd08-2096c1e71d2c" (UID: "086f5371-c7ea-49ba-bd08-2096c1e71d2c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.066231 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.066269 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f5371-c7ea-49ba-bd08-2096c1e71d2c-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.096622 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.134327 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62728408-1835-4e60-9d36-dac31093b2b9" path="/var/lib/kubelet/pods/62728408-1835-4e60-9d36-dac31093b2b9/volumes" Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.249388 4954 generic.go:334] "Generic (PLEG): container finished" podID="a767b973-ed53-4675-93a8-db722496f8d7" containerID="340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308" exitCode=0 Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.249835 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" event={"ID":"a767b973-ed53-4675-93a8-db722496f8d7","Type":"ContainerDied","Data":"340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308"} Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.250022 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" event={"ID":"a767b973-ed53-4675-93a8-db722496f8d7","Type":"ContainerStarted","Data":"fef072276e33a16173984619b7e235b22f7313fda5e2f16bebcaf81bd34824c3"} Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.254520 4954 generic.go:334] "Generic (PLEG): container finished" podID="6d66ceb0-8567-43c3-bbda-46693b51be82" containerID="c9d3f5a34ca8dd623b13c7dc430c3ca2de819e4ca4cf499e0b64eeeb5a13997c" exitCode=0 Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.254607 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" event={"ID":"6d66ceb0-8567-43c3-bbda-46693b51be82","Type":"ContainerDied","Data":"c9d3f5a34ca8dd623b13c7dc430c3ca2de819e4ca4cf499e0b64eeeb5a13997c"} Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.254639 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" event={"ID":"6d66ceb0-8567-43c3-bbda-46693b51be82","Type":"ContainerStarted","Data":"f80e40e8b9f3835184d2c6014ce5b9e19e6d6bab730ed3a5902764a203df0ddb"} Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.263582 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.266759 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-kr46n" event={"ID":"086f5371-c7ea-49ba-bd08-2096c1e71d2c","Type":"ContainerDied","Data":"4e93121d843d2d3bc26bac95e92f539bfd6f2d50225df499e789db0063a728d4"} Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.266826 4954 scope.go:117] "RemoveContainer" containerID="6e1f41f878eced48b084afec2fcdb5a0cc6c81e21a0b138ad2ed41fb3990acb6" Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.336605 4954 scope.go:117] "RemoveContainer" containerID="08b8b7079738285e018a76e393a3ae16abec97a138723bbcaada12a1b500fb5b" Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.351936 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kr46n"] Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.379755 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-kr46n"] Dec 09 17:25:58 crc kubenswrapper[4954]: I1209 17:25:58.583231 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 17:25:58 crc kubenswrapper[4954]: W1209 17:25:58.590116 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb77bf96d_5422_416f_ba9d_cd20e94d991e.slice/crio-acc952a38c6fd52695bd8458244e2cad788404315d8d195255b074b047f17448 WatchSource:0}: Error finding container acc952a38c6fd52695bd8458244e2cad788404315d8d195255b074b047f17448: Status 404 returned error can't find the container with id acc952a38c6fd52695bd8458244e2cad788404315d8d195255b074b047f17448 Dec 09 17:25:59 crc kubenswrapper[4954]: I1209 17:25:59.121272 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:25:59 crc kubenswrapper[4954]: E1209 17:25:59.122061 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:25:59 crc kubenswrapper[4954]: I1209 17:25:59.273748 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" event={"ID":"a767b973-ed53-4675-93a8-db722496f8d7","Type":"ContainerStarted","Data":"9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f"} Dec 09 17:25:59 crc kubenswrapper[4954]: I1209 17:25:59.273980 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:25:59 crc kubenswrapper[4954]: I1209 17:25:59.276103 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" event={"ID":"6d66ceb0-8567-43c3-bbda-46693b51be82","Type":"ContainerStarted","Data":"59f56ba85dc9ba7fff54d0d1a192fcf8dc99d8efd77c2f8f92cee8b2b0fc69d5"} Dec 09 17:25:59 crc kubenswrapper[4954]: I1209 17:25:59.276290 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:25:59 crc kubenswrapper[4954]: I1209 17:25:59.277242 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b77bf96d-5422-416f-ba9d-cd20e94d991e","Type":"ContainerStarted","Data":"acc952a38c6fd52695bd8458244e2cad788404315d8d195255b074b047f17448"} Dec 09 17:25:59 crc kubenswrapper[4954]: I1209 17:25:59.278518 4954 generic.go:334] "Generic (PLEG): container finished" podID="f0da1b73-3249-449b-86f9-12206ce93aab" containerID="bfa38e1cf3605a9d66bcda5a2cb2ed06a8bfbea9eb7f3287ed2e14d8d3509a38" exitCode=0 Dec 09 17:25:59 crc kubenswrapper[4954]: I1209 17:25:59.278603 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f0da1b73-3249-449b-86f9-12206ce93aab","Type":"ContainerDied","Data":"bfa38e1cf3605a9d66bcda5a2cb2ed06a8bfbea9eb7f3287ed2e14d8d3509a38"} Dec 09 17:25:59 crc kubenswrapper[4954]: I1209 17:25:59.314620 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" podStartSLOduration=5.314562759 podStartE2EDuration="5.314562759s" podCreationTimestamp="2025-12-09 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:25:59.308718316 +0000 UTC m=+1755.696892136" watchObservedRunningTime="2025-12-09 17:25:59.314562759 +0000 UTC m=+1755.702736579" Dec 09 17:25:59 crc kubenswrapper[4954]: I1209 17:25:59.338291 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" podStartSLOduration=5.33826854 podStartE2EDuration="5.33826854s" podCreationTimestamp="2025-12-09 17:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:25:59.336625669 +0000 UTC m=+1755.724799489" watchObservedRunningTime="2025-12-09 17:25:59.33826854 +0000 UTC m=+1755.726442360" Dec 09 17:26:00 crc kubenswrapper[4954]: I1209 17:26:00.155243 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="086f5371-c7ea-49ba-bd08-2096c1e71d2c" path="/var/lib/kubelet/pods/086f5371-c7ea-49ba-bd08-2096c1e71d2c/volumes" Dec 09 17:26:00 crc kubenswrapper[4954]: I1209 17:26:00.342783 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 09 17:26:01 crc kubenswrapper[4954]: I1209 17:26:01.300255 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b77bf96d-5422-416f-ba9d-cd20e94d991e","Type":"ContainerStarted","Data":"8186f5a2b7dfd0e883a7d456579efa41a55c7fc78908341abe63c6d904a2fa23"} Dec 09 17:26:01 crc kubenswrapper[4954]: I1209 17:26:01.300970 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 09 17:26:01 crc kubenswrapper[4954]: I1209 17:26:01.300987 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b77bf96d-5422-416f-ba9d-cd20e94d991e","Type":"ContainerStarted","Data":"bc481fe0e46c24e8f9013c3e7588fa7682c35fa6aecf54914849f853de237915"} Dec 09 17:26:03 crc kubenswrapper[4954]: I1209 17:26:03.554343 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 09 17:26:03 crc kubenswrapper[4954]: I1209 17:26:03.554787 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 09 17:26:03 crc kubenswrapper[4954]: I1209 17:26:03.637213 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 09 17:26:03 crc kubenswrapper[4954]: I1209 17:26:03.664572 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=4.781637032 podStartE2EDuration="6.664546686s" podCreationTimestamp="2025-12-09 17:25:57 +0000 UTC" firstStartedPulling="2025-12-09 17:25:58.593363992 +0000 UTC m=+1754.981537822" lastFinishedPulling="2025-12-09 17:26:00.476273656 +0000 UTC m=+1756.864447476" observedRunningTime="2025-12-09 17:26:01.325665897 +0000 UTC m=+1757.713839737" watchObservedRunningTime="2025-12-09 17:26:03.664546686 +0000 UTC m=+1760.052720506" Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.412900 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.624376 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.917947 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-bfd664845-r74kf" podUID="bb473662-281d-46be-996e-02370393136b" containerName="console" containerID="cri-o://41cd1fb4c7e3b2a0b444d6593701517406c2247bb6e0d30d677685170ca77192" gracePeriod=15 Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.936474 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cfdd-account-create-update-tqjv4"] Dec 09 17:26:04 crc kubenswrapper[4954]: E1209 17:26:04.939243 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="086f5371-c7ea-49ba-bd08-2096c1e71d2c" containerName="init" Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.939282 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="086f5371-c7ea-49ba-bd08-2096c1e71d2c" containerName="init" Dec 09 17:26:04 crc kubenswrapper[4954]: E1209 17:26:04.939302 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="086f5371-c7ea-49ba-bd08-2096c1e71d2c" containerName="dnsmasq-dns" Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.939310 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="086f5371-c7ea-49ba-bd08-2096c1e71d2c" containerName="dnsmasq-dns" Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.939618 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="086f5371-c7ea-49ba-bd08-2096c1e71d2c" containerName="dnsmasq-dns" Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.940751 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cfdd-account-create-update-tqjv4" Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.944191 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.948847 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cfdd-account-create-update-tqjv4"] Dec 09 17:26:04 crc kubenswrapper[4954]: I1209 17:26:04.999009 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-8grd9"] Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.000942 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8grd9" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.012666 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-8grd9"] Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.024236 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.024289 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.071357 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fldl8\" (UniqueName: \"kubernetes.io/projected/e6438189-ff16-4038-ba8d-3ee0972e45a3-kube-api-access-fldl8\") pod \"keystone-cfdd-account-create-update-tqjv4\" (UID: \"e6438189-ff16-4038-ba8d-3ee0972e45a3\") " pod="openstack/keystone-cfdd-account-create-update-tqjv4" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.071830 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6438189-ff16-4038-ba8d-3ee0972e45a3-operator-scripts\") pod \"keystone-cfdd-account-create-update-tqjv4\" (UID: \"e6438189-ff16-4038-ba8d-3ee0972e45a3\") " pod="openstack/keystone-cfdd-account-create-update-tqjv4" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.114788 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.180997 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-599bw"] Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.184399 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad0e11b4-6dcf-437b-8798-1acdacc534ed-operator-scripts\") pod \"keystone-db-create-8grd9\" (UID: \"ad0e11b4-6dcf-437b-8798-1acdacc534ed\") " pod="openstack/keystone-db-create-8grd9" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.185058 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fldl8\" (UniqueName: \"kubernetes.io/projected/e6438189-ff16-4038-ba8d-3ee0972e45a3-kube-api-access-fldl8\") pod \"keystone-cfdd-account-create-update-tqjv4\" (UID: \"e6438189-ff16-4038-ba8d-3ee0972e45a3\") " pod="openstack/keystone-cfdd-account-create-update-tqjv4" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.185152 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr7rj\" (UniqueName: \"kubernetes.io/projected/ad0e11b4-6dcf-437b-8798-1acdacc534ed-kube-api-access-rr7rj\") pod \"keystone-db-create-8grd9\" (UID: \"ad0e11b4-6dcf-437b-8798-1acdacc534ed\") " pod="openstack/keystone-db-create-8grd9" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.185185 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6438189-ff16-4038-ba8d-3ee0972e45a3-operator-scripts\") pod \"keystone-cfdd-account-create-update-tqjv4\" (UID: \"e6438189-ff16-4038-ba8d-3ee0972e45a3\") " pod="openstack/keystone-cfdd-account-create-update-tqjv4" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.186313 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6438189-ff16-4038-ba8d-3ee0972e45a3-operator-scripts\") pod \"keystone-cfdd-account-create-update-tqjv4\" (UID: \"e6438189-ff16-4038-ba8d-3ee0972e45a3\") " pod="openstack/keystone-cfdd-account-create-update-tqjv4" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.237065 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fldl8\" (UniqueName: \"kubernetes.io/projected/e6438189-ff16-4038-ba8d-3ee0972e45a3-kube-api-access-fldl8\") pod \"keystone-cfdd-account-create-update-tqjv4\" (UID: \"e6438189-ff16-4038-ba8d-3ee0972e45a3\") " pod="openstack/keystone-cfdd-account-create-update-tqjv4" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.287071 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr7rj\" (UniqueName: \"kubernetes.io/projected/ad0e11b4-6dcf-437b-8798-1acdacc534ed-kube-api-access-rr7rj\") pod \"keystone-db-create-8grd9\" (UID: \"ad0e11b4-6dcf-437b-8798-1acdacc534ed\") " pod="openstack/keystone-db-create-8grd9" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.287245 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad0e11b4-6dcf-437b-8798-1acdacc534ed-operator-scripts\") pod \"keystone-db-create-8grd9\" (UID: \"ad0e11b4-6dcf-437b-8798-1acdacc534ed\") " pod="openstack/keystone-db-create-8grd9" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.287503 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a366-account-create-update-zkvjq"] Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.289062 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a366-account-create-update-zkvjq" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.289176 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad0e11b4-6dcf-437b-8798-1acdacc534ed-operator-scripts\") pod \"keystone-db-create-8grd9\" (UID: \"ad0e11b4-6dcf-437b-8798-1acdacc534ed\") " pod="openstack/keystone-db-create-8grd9" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.298885 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-c8s82"] Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.300530 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c8s82" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.305903 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.309658 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a366-account-create-update-zkvjq"] Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.317785 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr7rj\" (UniqueName: \"kubernetes.io/projected/ad0e11b4-6dcf-437b-8798-1acdacc534ed-kube-api-access-rr7rj\") pod \"keystone-db-create-8grd9\" (UID: \"ad0e11b4-6dcf-437b-8798-1acdacc534ed\") " pod="openstack/keystone-db-create-8grd9" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.322429 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cfdd-account-create-update-tqjv4" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.342039 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c8s82"] Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.343217 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8grd9" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.384396 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-bfd664845-r74kf_bb473662-281d-46be-996e-02370393136b/console/0.log" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.384889 4954 generic.go:334] "Generic (PLEG): container finished" podID="bb473662-281d-46be-996e-02370393136b" containerID="41cd1fb4c7e3b2a0b444d6593701517406c2247bb6e0d30d677685170ca77192" exitCode=2 Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.385125 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" podUID="6d66ceb0-8567-43c3-bbda-46693b51be82" containerName="dnsmasq-dns" containerID="cri-o://59f56ba85dc9ba7fff54d0d1a192fcf8dc99d8efd77c2f8f92cee8b2b0fc69d5" gracePeriod=10 Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.385258 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-bfd664845-r74kf" event={"ID":"bb473662-281d-46be-996e-02370393136b","Type":"ContainerDied","Data":"41cd1fb4c7e3b2a0b444d6593701517406c2247bb6e0d30d677685170ca77192"} Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.388765 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hzmd\" (UniqueName: \"kubernetes.io/projected/56f3777f-f5d6-4a05-855e-8a28b67a07eb-kube-api-access-7hzmd\") pod \"placement-db-create-c8s82\" (UID: \"56f3777f-f5d6-4a05-855e-8a28b67a07eb\") " pod="openstack/placement-db-create-c8s82" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.388807 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56f3777f-f5d6-4a05-855e-8a28b67a07eb-operator-scripts\") pod \"placement-db-create-c8s82\" (UID: \"56f3777f-f5d6-4a05-855e-8a28b67a07eb\") " pod="openstack/placement-db-create-c8s82" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.388855 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13cbee2b-45bb-4d9e-8427-bb2574b8d284-operator-scripts\") pod \"placement-a366-account-create-update-zkvjq\" (UID: \"13cbee2b-45bb-4d9e-8427-bb2574b8d284\") " pod="openstack/placement-a366-account-create-update-zkvjq" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.388934 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnjw4\" (UniqueName: \"kubernetes.io/projected/13cbee2b-45bb-4d9e-8427-bb2574b8d284-kube-api-access-vnjw4\") pod \"placement-a366-account-create-update-zkvjq\" (UID: \"13cbee2b-45bb-4d9e-8427-bb2574b8d284\") " pod="openstack/placement-a366-account-create-update-zkvjq" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.490975 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnjw4\" (UniqueName: \"kubernetes.io/projected/13cbee2b-45bb-4d9e-8427-bb2574b8d284-kube-api-access-vnjw4\") pod \"placement-a366-account-create-update-zkvjq\" (UID: \"13cbee2b-45bb-4d9e-8427-bb2574b8d284\") " pod="openstack/placement-a366-account-create-update-zkvjq" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.491079 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hzmd\" (UniqueName: \"kubernetes.io/projected/56f3777f-f5d6-4a05-855e-8a28b67a07eb-kube-api-access-7hzmd\") pod \"placement-db-create-c8s82\" (UID: \"56f3777f-f5d6-4a05-855e-8a28b67a07eb\") " pod="openstack/placement-db-create-c8s82" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.491131 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56f3777f-f5d6-4a05-855e-8a28b67a07eb-operator-scripts\") pod \"placement-db-create-c8s82\" (UID: \"56f3777f-f5d6-4a05-855e-8a28b67a07eb\") " pod="openstack/placement-db-create-c8s82" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.491212 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13cbee2b-45bb-4d9e-8427-bb2574b8d284-operator-scripts\") pod \"placement-a366-account-create-update-zkvjq\" (UID: \"13cbee2b-45bb-4d9e-8427-bb2574b8d284\") " pod="openstack/placement-a366-account-create-update-zkvjq" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.492840 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56f3777f-f5d6-4a05-855e-8a28b67a07eb-operator-scripts\") pod \"placement-db-create-c8s82\" (UID: \"56f3777f-f5d6-4a05-855e-8a28b67a07eb\") " pod="openstack/placement-db-create-c8s82" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.493187 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13cbee2b-45bb-4d9e-8427-bb2574b8d284-operator-scripts\") pod \"placement-a366-account-create-update-zkvjq\" (UID: \"13cbee2b-45bb-4d9e-8427-bb2574b8d284\") " pod="openstack/placement-a366-account-create-update-zkvjq" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.512842 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnjw4\" (UniqueName: \"kubernetes.io/projected/13cbee2b-45bb-4d9e-8427-bb2574b8d284-kube-api-access-vnjw4\") pod \"placement-a366-account-create-update-zkvjq\" (UID: \"13cbee2b-45bb-4d9e-8427-bb2574b8d284\") " pod="openstack/placement-a366-account-create-update-zkvjq" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.512937 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hzmd\" (UniqueName: \"kubernetes.io/projected/56f3777f-f5d6-4a05-855e-8a28b67a07eb-kube-api-access-7hzmd\") pod \"placement-db-create-c8s82\" (UID: \"56f3777f-f5d6-4a05-855e-8a28b67a07eb\") " pod="openstack/placement-db-create-c8s82" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.680301 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a366-account-create-update-zkvjq" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.694334 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c8s82" Dec 09 17:26:05 crc kubenswrapper[4954]: I1209 17:26:05.920412 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 09 17:26:06 crc kubenswrapper[4954]: I1209 17:26:06.020532 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 09 17:26:06 crc kubenswrapper[4954]: I1209 17:26:06.396884 4954 generic.go:334] "Generic (PLEG): container finished" podID="6d66ceb0-8567-43c3-bbda-46693b51be82" containerID="59f56ba85dc9ba7fff54d0d1a192fcf8dc99d8efd77c2f8f92cee8b2b0fc69d5" exitCode=0 Dec 09 17:26:06 crc kubenswrapper[4954]: I1209 17:26:06.397354 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" event={"ID":"6d66ceb0-8567-43c3-bbda-46693b51be82","Type":"ContainerDied","Data":"59f56ba85dc9ba7fff54d0d1a192fcf8dc99d8efd77c2f8f92cee8b2b0fc69d5"} Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.709722 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-q85b8"] Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.715194 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.744052 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-q85b8"] Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.745119 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvvct\" (UniqueName: \"kubernetes.io/projected/7801abe0-4c3f-4251-8b29-e8ff3e11b580-kube-api-access-bvvct\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.745257 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-config\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.745336 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.745523 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-dns-svc\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.745973 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.840065 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-5w9lz"] Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.842042 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.847998 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.848093 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvvct\" (UniqueName: \"kubernetes.io/projected/7801abe0-4c3f-4251-8b29-e8ff3e11b580-kube-api-access-bvvct\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.848142 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-config\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.848159 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.848184 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-dns-svc\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.849343 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-dns-svc\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.849892 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.850753 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-config\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.851583 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.855784 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-5w9lz"] Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.885477 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvvct\" (UniqueName: \"kubernetes.io/projected/7801abe0-4c3f-4251-8b29-e8ff3e11b580-kube-api-access-bvvct\") pod \"dnsmasq-dns-698758b865-q85b8\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.950127 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-5w9lz\" (UID: \"3c2c10db-d15d-4fc4-b430-80b7cfb2c401\") " pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.950188 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr7sm\" (UniqueName: \"kubernetes.io/projected/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-kube-api-access-sr7sm\") pod \"mysqld-exporter-openstack-db-create-5w9lz\" (UID: \"3c2c10db-d15d-4fc4-b430-80b7cfb2c401\") " pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" Dec 09 17:26:07 crc kubenswrapper[4954]: I1209 17:26:07.956529 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0126-account-create-update-z7h68"] Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:07.961746 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:07.964884 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:07.967850 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0126-account-create-update-z7h68"] Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.048894 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.056097 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr7sm\" (UniqueName: \"kubernetes.io/projected/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-kube-api-access-sr7sm\") pod \"mysqld-exporter-openstack-db-create-5w9lz\" (UID: \"3c2c10db-d15d-4fc4-b430-80b7cfb2c401\") " pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.056305 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2vtw\" (UniqueName: \"kubernetes.io/projected/629b86ef-b3c8-445c-aab0-c59fb2359f57-kube-api-access-c2vtw\") pod \"mysqld-exporter-0126-account-create-update-z7h68\" (UID: \"629b86ef-b3c8-445c-aab0-c59fb2359f57\") " pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.056344 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/629b86ef-b3c8-445c-aab0-c59fb2359f57-operator-scripts\") pod \"mysqld-exporter-0126-account-create-update-z7h68\" (UID: \"629b86ef-b3c8-445c-aab0-c59fb2359f57\") " pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.056389 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-5w9lz\" (UID: \"3c2c10db-d15d-4fc4-b430-80b7cfb2c401\") " pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.057394 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-5w9lz\" (UID: \"3c2c10db-d15d-4fc4-b430-80b7cfb2c401\") " pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.083450 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr7sm\" (UniqueName: \"kubernetes.io/projected/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-kube-api-access-sr7sm\") pod \"mysqld-exporter-openstack-db-create-5w9lz\" (UID: \"3c2c10db-d15d-4fc4-b430-80b7cfb2c401\") " pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.158772 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2vtw\" (UniqueName: \"kubernetes.io/projected/629b86ef-b3c8-445c-aab0-c59fb2359f57-kube-api-access-c2vtw\") pod \"mysqld-exporter-0126-account-create-update-z7h68\" (UID: \"629b86ef-b3c8-445c-aab0-c59fb2359f57\") " pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.158843 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/629b86ef-b3c8-445c-aab0-c59fb2359f57-operator-scripts\") pod \"mysqld-exporter-0126-account-create-update-z7h68\" (UID: \"629b86ef-b3c8-445c-aab0-c59fb2359f57\") " pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.160314 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/629b86ef-b3c8-445c-aab0-c59fb2359f57-operator-scripts\") pod \"mysqld-exporter-0126-account-create-update-z7h68\" (UID: \"629b86ef-b3c8-445c-aab0-c59fb2359f57\") " pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.165730 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.180936 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2vtw\" (UniqueName: \"kubernetes.io/projected/629b86ef-b3c8-445c-aab0-c59fb2359f57-kube-api-access-c2vtw\") pod \"mysqld-exporter-0126-account-create-update-z7h68\" (UID: \"629b86ef-b3c8-445c-aab0-c59fb2359f57\") " pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.331214 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.417089 4954 generic.go:334] "Generic (PLEG): container finished" podID="28718e77-f5b5-4d67-afbc-c1157354bc47" containerID="fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad" exitCode=0 Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.417149 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"28718e77-f5b5-4d67-afbc-c1157354bc47","Type":"ContainerDied","Data":"fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad"} Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.877477 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.887765 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.891005 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-n25zw" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.891215 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.891291 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.891492 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 09 17:26:08 crc kubenswrapper[4954]: I1209 17:26:08.907266 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.080549 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/69b248e7-6cb4-4805-a01f-537fe58aa42d-cache\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.080919 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.081221 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfqxg\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-kube-api-access-sfqxg\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.081310 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/69b248e7-6cb4-4805-a01f-537fe58aa42d-lock\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.081393 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.184073 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfqxg\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-kube-api-access-sfqxg\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.184615 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/69b248e7-6cb4-4805-a01f-537fe58aa42d-lock\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.184657 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.184700 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/69b248e7-6cb4-4805-a01f-537fe58aa42d-cache\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.184827 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: E1209 17:26:09.184982 4954 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 17:26:09 crc kubenswrapper[4954]: E1209 17:26:09.185021 4954 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 17:26:09 crc kubenswrapper[4954]: E1209 17:26:09.185101 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift podName:69b248e7-6cb4-4805-a01f-537fe58aa42d nodeName:}" failed. No retries permitted until 2025-12-09 17:26:09.685074389 +0000 UTC m=+1766.073248199 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift") pod "swift-storage-0" (UID: "69b248e7-6cb4-4805-a01f-537fe58aa42d") : configmap "swift-ring-files" not found Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.185380 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.185434 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/69b248e7-6cb4-4805-a01f-537fe58aa42d-lock\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.185617 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/69b248e7-6cb4-4805-a01f-537fe58aa42d-cache\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.207511 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfqxg\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-kube-api-access-sfqxg\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.223321 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.419422 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-4vtb6"] Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.421233 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.424782 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.425061 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.425490 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.436290 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-4vtb6"] Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.593546 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-combined-ca-bundle\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.593846 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-dispersionconf\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.593977 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-swiftconf\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.594022 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-496j9\" (UniqueName: \"kubernetes.io/projected/9fc44917-47c6-450b-b393-bad021024fcd-kube-api-access-496j9\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.594054 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-ring-data-devices\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.594171 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-scripts\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.594260 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fc44917-47c6-450b-b393-bad021024fcd-etc-swift\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.622878 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" podUID="6d66ceb0-8567-43c3-bbda-46693b51be82" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.140:5353: connect: connection refused" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.695751 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-combined-ca-bundle\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.695824 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-dispersionconf\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.695896 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-swiftconf\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.695921 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-496j9\" (UniqueName: \"kubernetes.io/projected/9fc44917-47c6-450b-b393-bad021024fcd-kube-api-access-496j9\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.695943 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-ring-data-devices\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.695994 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.696030 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-scripts\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.696104 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fc44917-47c6-450b-b393-bad021024fcd-etc-swift\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: E1209 17:26:09.696360 4954 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 17:26:09 crc kubenswrapper[4954]: E1209 17:26:09.696391 4954 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 17:26:09 crc kubenswrapper[4954]: E1209 17:26:09.696449 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift podName:69b248e7-6cb4-4805-a01f-537fe58aa42d nodeName:}" failed. No retries permitted until 2025-12-09 17:26:10.696426668 +0000 UTC m=+1767.084600488 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift") pod "swift-storage-0" (UID: "69b248e7-6cb4-4805-a01f-537fe58aa42d") : configmap "swift-ring-files" not found Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.697237 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fc44917-47c6-450b-b393-bad021024fcd-etc-swift\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.697731 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-scripts\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.697763 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-ring-data-devices\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.702545 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-swiftconf\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.703668 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-dispersionconf\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.704712 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-combined-ca-bundle\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.721460 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-496j9\" (UniqueName: \"kubernetes.io/projected/9fc44917-47c6-450b-b393-bad021024fcd-kube-api-access-496j9\") pod \"swift-ring-rebalance-4vtb6\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:09 crc kubenswrapper[4954]: I1209 17:26:09.743287 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.509028 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-c4jlk"] Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.511209 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c4jlk" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.523902 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-c4jlk"] Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.616371 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f22851f-5367-4fc7-92b4-6114a4d0449a-operator-scripts\") pod \"glance-db-create-c4jlk\" (UID: \"6f22851f-5367-4fc7-92b4-6114a4d0449a\") " pod="openstack/glance-db-create-c4jlk" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.616437 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk7zr\" (UniqueName: \"kubernetes.io/projected/6f22851f-5367-4fc7-92b4-6114a4d0449a-kube-api-access-xk7zr\") pod \"glance-db-create-c4jlk\" (UID: \"6f22851f-5367-4fc7-92b4-6114a4d0449a\") " pod="openstack/glance-db-create-c4jlk" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.686035 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-01f1-account-create-update-zd7hl"] Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.692475 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-01f1-account-create-update-zd7hl" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.698201 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.723565 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:10 crc kubenswrapper[4954]: E1209 17:26:10.723717 4954 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 17:26:10 crc kubenswrapper[4954]: E1209 17:26:10.723753 4954 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 17:26:10 crc kubenswrapper[4954]: E1209 17:26:10.723808 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift podName:69b248e7-6cb4-4805-a01f-537fe58aa42d nodeName:}" failed. No retries permitted until 2025-12-09 17:26:12.723785609 +0000 UTC m=+1769.111959439 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift") pod "swift-storage-0" (UID: "69b248e7-6cb4-4805-a01f-537fe58aa42d") : configmap "swift-ring-files" not found Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.724080 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f22851f-5367-4fc7-92b4-6114a4d0449a-operator-scripts\") pod \"glance-db-create-c4jlk\" (UID: \"6f22851f-5367-4fc7-92b4-6114a4d0449a\") " pod="openstack/glance-db-create-c4jlk" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.724185 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk7zr\" (UniqueName: \"kubernetes.io/projected/6f22851f-5367-4fc7-92b4-6114a4d0449a-kube-api-access-xk7zr\") pod \"glance-db-create-c4jlk\" (UID: \"6f22851f-5367-4fc7-92b4-6114a4d0449a\") " pod="openstack/glance-db-create-c4jlk" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.725333 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f22851f-5367-4fc7-92b4-6114a4d0449a-operator-scripts\") pod \"glance-db-create-c4jlk\" (UID: \"6f22851f-5367-4fc7-92b4-6114a4d0449a\") " pod="openstack/glance-db-create-c4jlk" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.731392 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-01f1-account-create-update-zd7hl"] Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.750093 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk7zr\" (UniqueName: \"kubernetes.io/projected/6f22851f-5367-4fc7-92b4-6114a4d0449a-kube-api-access-xk7zr\") pod \"glance-db-create-c4jlk\" (UID: \"6f22851f-5367-4fc7-92b4-6114a4d0449a\") " pod="openstack/glance-db-create-c4jlk" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.787111 4954 patch_prober.go:28] interesting pod/console-bfd664845-r74kf container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.87:8443/health\": dial tcp 10.217.0.87:8443: connect: connection refused" start-of-body= Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.787190 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-bfd664845-r74kf" podUID="bb473662-281d-46be-996e-02370393136b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.87:8443/health\": dial tcp 10.217.0.87:8443: connect: connection refused" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.837132 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4f98d10-48e9-4d4f-abde-aef5db23a39e-operator-scripts\") pod \"glance-01f1-account-create-update-zd7hl\" (UID: \"f4f98d10-48e9-4d4f-abde-aef5db23a39e\") " pod="openstack/glance-01f1-account-create-update-zd7hl" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.837262 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdk2m\" (UniqueName: \"kubernetes.io/projected/f4f98d10-48e9-4d4f-abde-aef5db23a39e-kube-api-access-pdk2m\") pod \"glance-01f1-account-create-update-zd7hl\" (UID: \"f4f98d10-48e9-4d4f-abde-aef5db23a39e\") " pod="openstack/glance-01f1-account-create-update-zd7hl" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.862284 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c4jlk" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.942757 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4f98d10-48e9-4d4f-abde-aef5db23a39e-operator-scripts\") pod \"glance-01f1-account-create-update-zd7hl\" (UID: \"f4f98d10-48e9-4d4f-abde-aef5db23a39e\") " pod="openstack/glance-01f1-account-create-update-zd7hl" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.942872 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdk2m\" (UniqueName: \"kubernetes.io/projected/f4f98d10-48e9-4d4f-abde-aef5db23a39e-kube-api-access-pdk2m\") pod \"glance-01f1-account-create-update-zd7hl\" (UID: \"f4f98d10-48e9-4d4f-abde-aef5db23a39e\") " pod="openstack/glance-01f1-account-create-update-zd7hl" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.943559 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4f98d10-48e9-4d4f-abde-aef5db23a39e-operator-scripts\") pod \"glance-01f1-account-create-update-zd7hl\" (UID: \"f4f98d10-48e9-4d4f-abde-aef5db23a39e\") " pod="openstack/glance-01f1-account-create-update-zd7hl" Dec 09 17:26:10 crc kubenswrapper[4954]: I1209 17:26:10.970850 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdk2m\" (UniqueName: \"kubernetes.io/projected/f4f98d10-48e9-4d4f-abde-aef5db23a39e-kube-api-access-pdk2m\") pod \"glance-01f1-account-create-update-zd7hl\" (UID: \"f4f98d10-48e9-4d4f-abde-aef5db23a39e\") " pod="openstack/glance-01f1-account-create-update-zd7hl" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.121793 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:26:11 crc kubenswrapper[4954]: E1209 17:26:11.122497 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.153422 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-01f1-account-create-update-zd7hl" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.154811 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.253512 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-dns-svc\") pod \"6d66ceb0-8567-43c3-bbda-46693b51be82\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.253794 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-ovsdbserver-nb\") pod \"6d66ceb0-8567-43c3-bbda-46693b51be82\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.253834 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-config\") pod \"6d66ceb0-8567-43c3-bbda-46693b51be82\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.254994 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w92l2\" (UniqueName: \"kubernetes.io/projected/6d66ceb0-8567-43c3-bbda-46693b51be82-kube-api-access-w92l2\") pod \"6d66ceb0-8567-43c3-bbda-46693b51be82\" (UID: \"6d66ceb0-8567-43c3-bbda-46693b51be82\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.266141 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d66ceb0-8567-43c3-bbda-46693b51be82-kube-api-access-w92l2" (OuterVolumeSpecName: "kube-api-access-w92l2") pod "6d66ceb0-8567-43c3-bbda-46693b51be82" (UID: "6d66ceb0-8567-43c3-bbda-46693b51be82"). InnerVolumeSpecName "kube-api-access-w92l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.280731 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w92l2\" (UniqueName: \"kubernetes.io/projected/6d66ceb0-8567-43c3-bbda-46693b51be82-kube-api-access-w92l2\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.388184 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6d66ceb0-8567-43c3-bbda-46693b51be82" (UID: "6d66ceb0-8567-43c3-bbda-46693b51be82"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.391384 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-config" (OuterVolumeSpecName: "config") pod "6d66ceb0-8567-43c3-bbda-46693b51be82" (UID: "6d66ceb0-8567-43c3-bbda-46693b51be82"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.413406 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6d66ceb0-8567-43c3-bbda-46693b51be82" (UID: "6d66ceb0-8567-43c3-bbda-46693b51be82"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.486229 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-bfd664845-r74kf_bb473662-281d-46be-996e-02370393136b/console/0.log" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.486543 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.488364 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.488387 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.488397 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d66ceb0-8567-43c3-bbda-46693b51be82-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.526170 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-bfd664845-r74kf_bb473662-281d-46be-996e-02370393136b/console/0.log" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.526368 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-bfd664845-r74kf" event={"ID":"bb473662-281d-46be-996e-02370393136b","Type":"ContainerDied","Data":"6bd945089dc93f02a3d79e9494098d989a604a78f2067273b6f6d3d26b955889"} Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.526423 4954 scope.go:117] "RemoveContainer" containerID="41cd1fb4c7e3b2a0b444d6593701517406c2247bb6e0d30d677685170ca77192" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.526412 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-bfd664845-r74kf" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.568796 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" event={"ID":"6d66ceb0-8567-43c3-bbda-46693b51be82","Type":"ContainerDied","Data":"f80e40e8b9f3835184d2c6014ce5b9e19e6d6bab730ed3a5902764a203df0ddb"} Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.568911 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-599bw" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.589484 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-service-ca\") pod \"bb473662-281d-46be-996e-02370393136b\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.589559 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-serving-cert\") pod \"bb473662-281d-46be-996e-02370393136b\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.589612 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-oauth-config\") pod \"bb473662-281d-46be-996e-02370393136b\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.589671 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-oauth-serving-cert\") pod \"bb473662-281d-46be-996e-02370393136b\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.589725 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-trusted-ca-bundle\") pod \"bb473662-281d-46be-996e-02370393136b\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.589818 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-console-config\") pod \"bb473662-281d-46be-996e-02370393136b\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.589894 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrs6p\" (UniqueName: \"kubernetes.io/projected/bb473662-281d-46be-996e-02370393136b-kube-api-access-lrs6p\") pod \"bb473662-281d-46be-996e-02370393136b\" (UID: \"bb473662-281d-46be-996e-02370393136b\") " Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.590774 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "bb473662-281d-46be-996e-02370393136b" (UID: "bb473662-281d-46be-996e-02370393136b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.591380 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-service-ca" (OuterVolumeSpecName: "service-ca") pod "bb473662-281d-46be-996e-02370393136b" (UID: "bb473662-281d-46be-996e-02370393136b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.591824 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "bb473662-281d-46be-996e-02370393136b" (UID: "bb473662-281d-46be-996e-02370393136b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.592086 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-console-config" (OuterVolumeSpecName: "console-config") pod "bb473662-281d-46be-996e-02370393136b" (UID: "bb473662-281d-46be-996e-02370393136b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.595767 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "bb473662-281d-46be-996e-02370393136b" (UID: "bb473662-281d-46be-996e-02370393136b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.652326 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "bb473662-281d-46be-996e-02370393136b" (UID: "bb473662-281d-46be-996e-02370393136b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.663067 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb473662-281d-46be-996e-02370393136b-kube-api-access-lrs6p" (OuterVolumeSpecName: "kube-api-access-lrs6p") pod "bb473662-281d-46be-996e-02370393136b" (UID: "bb473662-281d-46be-996e-02370393136b"). InnerVolumeSpecName "kube-api-access-lrs6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.695245 4954 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.695873 4954 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.695963 4954 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.696065 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrs6p\" (UniqueName: \"kubernetes.io/projected/bb473662-281d-46be-996e-02370393136b-kube-api-access-lrs6p\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.696254 4954 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb473662-281d-46be-996e-02370393136b-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.696344 4954 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.696398 4954 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb473662-281d-46be-996e-02370393136b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.894543 4954 scope.go:117] "RemoveContainer" containerID="59f56ba85dc9ba7fff54d0d1a192fcf8dc99d8efd77c2f8f92cee8b2b0fc69d5" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.948737 4954 scope.go:117] "RemoveContainer" containerID="c9d3f5a34ca8dd623b13c7dc430c3ca2de819e4ca4cf499e0b64eeeb5a13997c" Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.951769 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-599bw"] Dec 09 17:26:11 crc kubenswrapper[4954]: I1209 17:26:11.978258 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-599bw"] Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.047679 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-bfd664845-r74kf"] Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.057956 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-bfd664845-r74kf"] Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.184361 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d66ceb0-8567-43c3-bbda-46693b51be82" path="/var/lib/kubelet/pods/6d66ceb0-8567-43c3-bbda-46693b51be82/volumes" Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.185433 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb473662-281d-46be-996e-02370393136b" path="/var/lib/kubelet/pods/bb473662-281d-46be-996e-02370393136b/volumes" Dec 09 17:26:12 crc kubenswrapper[4954]: E1209 17:26:12.334603 4954 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb473662_281d_46be_996e_02370393136b.slice/crio-6bd945089dc93f02a3d79e9494098d989a604a78f2067273b6f6d3d26b955889\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb473662_281d_46be_996e_02370393136b.slice\": RecentStats: unable to find data in memory cache]" Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.593196 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f0da1b73-3249-449b-86f9-12206ce93aab","Type":"ContainerStarted","Data":"d691d5676e0e085b4b4d4e79b096c31307d777d05ac20dc56d385175663416e9"} Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.597633 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"28718e77-f5b5-4d67-afbc-c1157354bc47","Type":"ContainerStarted","Data":"343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb"} Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.598060 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.633799 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=48.234678952 podStartE2EDuration="1m2.633772532s" podCreationTimestamp="2025-12-09 17:25:10 +0000 UTC" firstStartedPulling="2025-12-09 17:25:19.416160601 +0000 UTC m=+1715.804334421" lastFinishedPulling="2025-12-09 17:25:33.815254181 +0000 UTC m=+1730.203428001" observedRunningTime="2025-12-09 17:26:12.624399489 +0000 UTC m=+1769.012573309" watchObservedRunningTime="2025-12-09 17:26:12.633772532 +0000 UTC m=+1769.021946352" Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.722957 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-5w9lz"] Dec 09 17:26:12 crc kubenswrapper[4954]: W1209 17:26:12.737831 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7801abe0_4c3f_4251_8b29_e8ff3e11b580.slice/crio-7796d7e5f5ca2ff847c3a82e3ef6f44f70d21d3762d612de78ef6c24ecefb29a WatchSource:0}: Error finding container 7796d7e5f5ca2ff847c3a82e3ef6f44f70d21d3762d612de78ef6c24ecefb29a: Status 404 returned error can't find the container with id 7796d7e5f5ca2ff847c3a82e3ef6f44f70d21d3762d612de78ef6c24ecefb29a Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.740004 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-q85b8"] Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.848927 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:12 crc kubenswrapper[4954]: E1209 17:26:12.850386 4954 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 17:26:12 crc kubenswrapper[4954]: E1209 17:26:12.850414 4954 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 17:26:12 crc kubenswrapper[4954]: E1209 17:26:12.850492 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift podName:69b248e7-6cb4-4805-a01f-537fe58aa42d nodeName:}" failed. No retries permitted until 2025-12-09 17:26:16.850474046 +0000 UTC m=+1773.238647866 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift") pod "swift-storage-0" (UID: "69b248e7-6cb4-4805-a01f-537fe58aa42d") : configmap "swift-ring-files" not found Dec 09 17:26:12 crc kubenswrapper[4954]: I1209 17:26:12.878718 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cfdd-account-create-update-tqjv4"] Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.116762 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-4vtb6"] Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.130761 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0126-account-create-update-z7h68"] Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.137369 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a366-account-create-update-zkvjq"] Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.145134 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c8s82"] Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.172921 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-01f1-account-create-update-zd7hl"] Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.182149 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-c4jlk"] Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.192232 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-8grd9"] Dec 09 17:26:13 crc kubenswrapper[4954]: W1209 17:26:13.433836 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fc44917_47c6_450b_b393_bad021024fcd.slice/crio-9b8d18825561257eb7a7f77963ab63c8e59ecf69a4e7496fb1077a4c35dcf52f WatchSource:0}: Error finding container 9b8d18825561257eb7a7f77963ab63c8e59ecf69a4e7496fb1077a4c35dcf52f: Status 404 returned error can't find the container with id 9b8d18825561257eb7a7f77963ab63c8e59ecf69a4e7496fb1077a4c35dcf52f Dec 09 17:26:13 crc kubenswrapper[4954]: W1209 17:26:13.436359 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13cbee2b_45bb_4d9e_8427_bb2574b8d284.slice/crio-7ebfa47bcec03ede20586aeab6282e41766775f1e5d6f90027f9dfd23e20f692 WatchSource:0}: Error finding container 7ebfa47bcec03ede20586aeab6282e41766775f1e5d6f90027f9dfd23e20f692: Status 404 returned error can't find the container with id 7ebfa47bcec03ede20586aeab6282e41766775f1e5d6f90027f9dfd23e20f692 Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.606965 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" event={"ID":"629b86ef-b3c8-445c-aab0-c59fb2359f57","Type":"ContainerStarted","Data":"9aeaf034113fc9f19026fd66f5e4e5884b17b2d5af026b7c7e7ac73923d88958"} Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.608616 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-q85b8" event={"ID":"7801abe0-4c3f-4251-8b29-e8ff3e11b580","Type":"ContainerStarted","Data":"7796d7e5f5ca2ff847c3a82e3ef6f44f70d21d3762d612de78ef6c24ecefb29a"} Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.609826 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cfdd-account-create-update-tqjv4" event={"ID":"e6438189-ff16-4038-ba8d-3ee0972e45a3","Type":"ContainerStarted","Data":"efc61e3c6e70568469464b7484ce696bc18dcb14ec60fbd570128de34989a7ab"} Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.610926 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" event={"ID":"3c2c10db-d15d-4fc4-b430-80b7cfb2c401","Type":"ContainerStarted","Data":"39ac039d13ac8d51a5fcfd8716a8777b47d72edfdd2c6e508d04a03053edcf47"} Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.614939 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4vtb6" event={"ID":"9fc44917-47c6-450b-b393-bad021024fcd","Type":"ContainerStarted","Data":"9b8d18825561257eb7a7f77963ab63c8e59ecf69a4e7496fb1077a4c35dcf52f"} Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.616154 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a366-account-create-update-zkvjq" event={"ID":"13cbee2b-45bb-4d9e-8427-bb2574b8d284","Type":"ContainerStarted","Data":"7ebfa47bcec03ede20586aeab6282e41766775f1e5d6f90027f9dfd23e20f692"} Dec 09 17:26:13 crc kubenswrapper[4954]: W1209 17:26:13.730327 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f98d10_48e9_4d4f_abde_aef5db23a39e.slice/crio-0821a0e84b1854e2a50f7695cf1595220b2d3e07317ad8c561a1ba6fb7c62d1c WatchSource:0}: Error finding container 0821a0e84b1854e2a50f7695cf1595220b2d3e07317ad8c561a1ba6fb7c62d1c: Status 404 returned error can't find the container with id 0821a0e84b1854e2a50f7695cf1595220b2d3e07317ad8c561a1ba6fb7c62d1c Dec 09 17:26:13 crc kubenswrapper[4954]: I1209 17:26:13.874705 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.630235 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c4jlk" event={"ID":"6f22851f-5367-4fc7-92b4-6114a4d0449a","Type":"ContainerStarted","Data":"ce5b946d2591077b1daed5f36d9e99e25a3527cf293ef77ee6dc747fb18ebf7b"} Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.631110 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c4jlk" event={"ID":"6f22851f-5367-4fc7-92b4-6114a4d0449a","Type":"ContainerStarted","Data":"ed1068bf9ea9f65b955ac555a92eef385f4ec2f21629980060c99307fd2b58b9"} Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.634902 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8grd9" event={"ID":"ad0e11b4-6dcf-437b-8798-1acdacc534ed","Type":"ContainerStarted","Data":"cb0010c4e2f2f22c9accf9a84fe0a7fbb6c8a452ca2e1ef98edfb077874d7679"} Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.634957 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8grd9" event={"ID":"ad0e11b4-6dcf-437b-8798-1acdacc534ed","Type":"ContainerStarted","Data":"7f2c1c53aee0c8342e271ed52afd33b992e19106eeb559c786a2d2b993261957"} Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.644102 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cfdd-account-create-update-tqjv4" event={"ID":"e6438189-ff16-4038-ba8d-3ee0972e45a3","Type":"ContainerStarted","Data":"81dbcbdd15ecbd20c8cecc3dc608ef732a88b95f9097ed37cb3f6be09ef975a6"} Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.646753 4954 generic.go:334] "Generic (PLEG): container finished" podID="3c2c10db-d15d-4fc4-b430-80b7cfb2c401" containerID="cca583698e0b90d3649182f307906fd96b34a47fdc2eef0599292f61ac2a559a" exitCode=0 Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.646897 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" event={"ID":"3c2c10db-d15d-4fc4-b430-80b7cfb2c401","Type":"ContainerDied","Data":"cca583698e0b90d3649182f307906fd96b34a47fdc2eef0599292f61ac2a559a"} Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.652866 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c8s82" event={"ID":"56f3777f-f5d6-4a05-855e-8a28b67a07eb","Type":"ContainerStarted","Data":"7be2b29b530184976b6cd1da628b57dd1180642c1625978079fe5893c6d7ca0e"} Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.655457 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-01f1-account-create-update-zd7hl" event={"ID":"f4f98d10-48e9-4d4f-abde-aef5db23a39e","Type":"ContainerStarted","Data":"0821a0e84b1854e2a50f7695cf1595220b2d3e07317ad8c561a1ba6fb7c62d1c"} Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.659632 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-c4jlk" podStartSLOduration=4.659602912 podStartE2EDuration="4.659602912s" podCreationTimestamp="2025-12-09 17:26:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:14.647499403 +0000 UTC m=+1771.035673223" watchObservedRunningTime="2025-12-09 17:26:14.659602912 +0000 UTC m=+1771.047776732" Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.672877 4954 generic.go:334] "Generic (PLEG): container finished" podID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerID="bb0cef67501194586516c8f80b032e6a3bbfae73640e71f342fd1ac48279bdaa" exitCode=0 Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.672937 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-q85b8" event={"ID":"7801abe0-4c3f-4251-8b29-e8ff3e11b580","Type":"ContainerDied","Data":"bb0cef67501194586516c8f80b032e6a3bbfae73640e71f342fd1ac48279bdaa"} Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.767465 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cfdd-account-create-update-tqjv4" podStartSLOduration=10.767440618 podStartE2EDuration="10.767440618s" podCreationTimestamp="2025-12-09 17:26:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:14.756794195 +0000 UTC m=+1771.144968035" watchObservedRunningTime="2025-12-09 17:26:14.767440618 +0000 UTC m=+1771.155614428" Dec 09 17:26:14 crc kubenswrapper[4954]: I1209 17:26:14.836130 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-8grd9" podStartSLOduration=10.836101648 podStartE2EDuration="10.836101648s" podCreationTimestamp="2025-12-09 17:26:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:14.811381484 +0000 UTC m=+1771.199555304" watchObservedRunningTime="2025-12-09 17:26:14.836101648 +0000 UTC m=+1771.224275468" Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.687781 4954 generic.go:334] "Generic (PLEG): container finished" podID="ad0e11b4-6dcf-437b-8798-1acdacc534ed" containerID="cb0010c4e2f2f22c9accf9a84fe0a7fbb6c8a452ca2e1ef98edfb077874d7679" exitCode=0 Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.687906 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8grd9" event={"ID":"ad0e11b4-6dcf-437b-8798-1acdacc534ed","Type":"ContainerDied","Data":"cb0010c4e2f2f22c9accf9a84fe0a7fbb6c8a452ca2e1ef98edfb077874d7679"} Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.695745 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f0da1b73-3249-449b-86f9-12206ce93aab","Type":"ContainerStarted","Data":"8c9caea0a581dfbe0d2b6c0035f2f0087c3e7b00bb3b4c12dd285ab0ebe0e6b5"} Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.700920 4954 generic.go:334] "Generic (PLEG): container finished" podID="e6438189-ff16-4038-ba8d-3ee0972e45a3" containerID="81dbcbdd15ecbd20c8cecc3dc608ef732a88b95f9097ed37cb3f6be09ef975a6" exitCode=0 Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.701027 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cfdd-account-create-update-tqjv4" event={"ID":"e6438189-ff16-4038-ba8d-3ee0972e45a3","Type":"ContainerDied","Data":"81dbcbdd15ecbd20c8cecc3dc608ef732a88b95f9097ed37cb3f6be09ef975a6"} Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.708283 4954 generic.go:334] "Generic (PLEG): container finished" podID="56f3777f-f5d6-4a05-855e-8a28b67a07eb" containerID="550c11f7d963f566535d572ec244a4e3166b6b834be5d43a92cc18acd976a737" exitCode=0 Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.708347 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c8s82" event={"ID":"56f3777f-f5d6-4a05-855e-8a28b67a07eb","Type":"ContainerDied","Data":"550c11f7d963f566535d572ec244a4e3166b6b834be5d43a92cc18acd976a737"} Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.711459 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a366-account-create-update-zkvjq" event={"ID":"13cbee2b-45bb-4d9e-8427-bb2574b8d284","Type":"ContainerStarted","Data":"667162aaff42650f3f21cd35f3d7d9f258df7e8210d1932a36d2c2ecf9ce04bd"} Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.715725 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-01f1-account-create-update-zd7hl" event={"ID":"f4f98d10-48e9-4d4f-abde-aef5db23a39e","Type":"ContainerStarted","Data":"bd85f23b3130bdeb6dbfa020ea1e347b46593c4e137ef77a95d4897eba77da80"} Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.719230 4954 generic.go:334] "Generic (PLEG): container finished" podID="629b86ef-b3c8-445c-aab0-c59fb2359f57" containerID="1c7e0858f76d7fed7b2b05e3a15cea9ed4f931302180a0ed8bfffc3879ada9d5" exitCode=0 Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.719488 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" event={"ID":"629b86ef-b3c8-445c-aab0-c59fb2359f57","Type":"ContainerDied","Data":"1c7e0858f76d7fed7b2b05e3a15cea9ed4f931302180a0ed8bfffc3879ada9d5"} Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.722479 4954 generic.go:334] "Generic (PLEG): container finished" podID="6f22851f-5367-4fc7-92b4-6114a4d0449a" containerID="ce5b946d2591077b1daed5f36d9e99e25a3527cf293ef77ee6dc747fb18ebf7b" exitCode=0 Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.722503 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c4jlk" event={"ID":"6f22851f-5367-4fc7-92b4-6114a4d0449a","Type":"ContainerDied","Data":"ce5b946d2591077b1daed5f36d9e99e25a3527cf293ef77ee6dc747fb18ebf7b"} Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.801329 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-a366-account-create-update-zkvjq" podStartSLOduration=10.801300273 podStartE2EDuration="10.801300273s" podCreationTimestamp="2025-12-09 17:26:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:15.7887402 +0000 UTC m=+1772.176914030" watchObservedRunningTime="2025-12-09 17:26:15.801300273 +0000 UTC m=+1772.189474093" Dec 09 17:26:15 crc kubenswrapper[4954]: I1209 17:26:15.855631 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-01f1-account-create-update-zd7hl" podStartSLOduration=5.855586243 podStartE2EDuration="5.855586243s" podCreationTimestamp="2025-12-09 17:26:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:15.85452626 +0000 UTC m=+1772.242700080" watchObservedRunningTime="2025-12-09 17:26:15.855586243 +0000 UTC m=+1772.243760063" Dec 09 17:26:16 crc kubenswrapper[4954]: I1209 17:26:16.744811 4954 generic.go:334] "Generic (PLEG): container finished" podID="13cbee2b-45bb-4d9e-8427-bb2574b8d284" containerID="667162aaff42650f3f21cd35f3d7d9f258df7e8210d1932a36d2c2ecf9ce04bd" exitCode=0 Dec 09 17:26:16 crc kubenswrapper[4954]: I1209 17:26:16.745082 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a366-account-create-update-zkvjq" event={"ID":"13cbee2b-45bb-4d9e-8427-bb2574b8d284","Type":"ContainerDied","Data":"667162aaff42650f3f21cd35f3d7d9f258df7e8210d1932a36d2c2ecf9ce04bd"} Dec 09 17:26:16 crc kubenswrapper[4954]: I1209 17:26:16.747958 4954 generic.go:334] "Generic (PLEG): container finished" podID="f4f98d10-48e9-4d4f-abde-aef5db23a39e" containerID="bd85f23b3130bdeb6dbfa020ea1e347b46593c4e137ef77a95d4897eba77da80" exitCode=0 Dec 09 17:26:16 crc kubenswrapper[4954]: I1209 17:26:16.747999 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-01f1-account-create-update-zd7hl" event={"ID":"f4f98d10-48e9-4d4f-abde-aef5db23a39e","Type":"ContainerDied","Data":"bd85f23b3130bdeb6dbfa020ea1e347b46593c4e137ef77a95d4897eba77da80"} Dec 09 17:26:16 crc kubenswrapper[4954]: I1209 17:26:16.753413 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-q85b8" event={"ID":"7801abe0-4c3f-4251-8b29-e8ff3e11b580","Type":"ContainerStarted","Data":"36b7c152baf598616164b052b73516434a385be71cac9598195a5390df8eb505"} Dec 09 17:26:16 crc kubenswrapper[4954]: I1209 17:26:16.754026 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:16 crc kubenswrapper[4954]: I1209 17:26:16.800078 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-q85b8" podStartSLOduration=9.8000552 podStartE2EDuration="9.8000552s" podCreationTimestamp="2025-12-09 17:26:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:16.793088612 +0000 UTC m=+1773.181262432" watchObservedRunningTime="2025-12-09 17:26:16.8000552 +0000 UTC m=+1773.188229020" Dec 09 17:26:16 crc kubenswrapper[4954]: I1209 17:26:16.866733 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:16 crc kubenswrapper[4954]: E1209 17:26:16.867001 4954 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 17:26:16 crc kubenswrapper[4954]: E1209 17:26:16.867034 4954 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 17:26:16 crc kubenswrapper[4954]: E1209 17:26:16.867111 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift podName:69b248e7-6cb4-4805-a01f-537fe58aa42d nodeName:}" failed. No retries permitted until 2025-12-09 17:26:24.867086318 +0000 UTC m=+1781.255260138 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift") pod "swift-storage-0" (UID: "69b248e7-6cb4-4805-a01f-537fe58aa42d") : configmap "swift-ring-files" not found Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.573115 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c8s82" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.603949 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cfdd-account-create-update-tqjv4" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.610069 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.629621 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8grd9" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.639430 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.648849 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-01f1-account-create-update-zd7hl" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.663253 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a366-account-create-update-zkvjq" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.687475 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c4jlk" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696154 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56f3777f-f5d6-4a05-855e-8a28b67a07eb-operator-scripts\") pod \"56f3777f-f5d6-4a05-855e-8a28b67a07eb\" (UID: \"56f3777f-f5d6-4a05-855e-8a28b67a07eb\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696281 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdk2m\" (UniqueName: \"kubernetes.io/projected/f4f98d10-48e9-4d4f-abde-aef5db23a39e-kube-api-access-pdk2m\") pod \"f4f98d10-48e9-4d4f-abde-aef5db23a39e\" (UID: \"f4f98d10-48e9-4d4f-abde-aef5db23a39e\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696310 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr7rj\" (UniqueName: \"kubernetes.io/projected/ad0e11b4-6dcf-437b-8798-1acdacc534ed-kube-api-access-rr7rj\") pod \"ad0e11b4-6dcf-437b-8798-1acdacc534ed\" (UID: \"ad0e11b4-6dcf-437b-8798-1acdacc534ed\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696369 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13cbee2b-45bb-4d9e-8427-bb2574b8d284-operator-scripts\") pod \"13cbee2b-45bb-4d9e-8427-bb2574b8d284\" (UID: \"13cbee2b-45bb-4d9e-8427-bb2574b8d284\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696426 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fldl8\" (UniqueName: \"kubernetes.io/projected/e6438189-ff16-4038-ba8d-3ee0972e45a3-kube-api-access-fldl8\") pod \"e6438189-ff16-4038-ba8d-3ee0972e45a3\" (UID: \"e6438189-ff16-4038-ba8d-3ee0972e45a3\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696460 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnjw4\" (UniqueName: \"kubernetes.io/projected/13cbee2b-45bb-4d9e-8427-bb2574b8d284-kube-api-access-vnjw4\") pod \"13cbee2b-45bb-4d9e-8427-bb2574b8d284\" (UID: \"13cbee2b-45bb-4d9e-8427-bb2574b8d284\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696551 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-operator-scripts\") pod \"3c2c10db-d15d-4fc4-b430-80b7cfb2c401\" (UID: \"3c2c10db-d15d-4fc4-b430-80b7cfb2c401\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696581 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4f98d10-48e9-4d4f-abde-aef5db23a39e-operator-scripts\") pod \"f4f98d10-48e9-4d4f-abde-aef5db23a39e\" (UID: \"f4f98d10-48e9-4d4f-abde-aef5db23a39e\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696688 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hzmd\" (UniqueName: \"kubernetes.io/projected/56f3777f-f5d6-4a05-855e-8a28b67a07eb-kube-api-access-7hzmd\") pod \"56f3777f-f5d6-4a05-855e-8a28b67a07eb\" (UID: \"56f3777f-f5d6-4a05-855e-8a28b67a07eb\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696747 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad0e11b4-6dcf-437b-8798-1acdacc534ed-operator-scripts\") pod \"ad0e11b4-6dcf-437b-8798-1acdacc534ed\" (UID: \"ad0e11b4-6dcf-437b-8798-1acdacc534ed\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696768 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/629b86ef-b3c8-445c-aab0-c59fb2359f57-operator-scripts\") pod \"629b86ef-b3c8-445c-aab0-c59fb2359f57\" (UID: \"629b86ef-b3c8-445c-aab0-c59fb2359f57\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696801 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr7sm\" (UniqueName: \"kubernetes.io/projected/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-kube-api-access-sr7sm\") pod \"3c2c10db-d15d-4fc4-b430-80b7cfb2c401\" (UID: \"3c2c10db-d15d-4fc4-b430-80b7cfb2c401\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696824 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2vtw\" (UniqueName: \"kubernetes.io/projected/629b86ef-b3c8-445c-aab0-c59fb2359f57-kube-api-access-c2vtw\") pod \"629b86ef-b3c8-445c-aab0-c59fb2359f57\" (UID: \"629b86ef-b3c8-445c-aab0-c59fb2359f57\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.696886 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6438189-ff16-4038-ba8d-3ee0972e45a3-operator-scripts\") pod \"e6438189-ff16-4038-ba8d-3ee0972e45a3\" (UID: \"e6438189-ff16-4038-ba8d-3ee0972e45a3\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.700070 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6438189-ff16-4038-ba8d-3ee0972e45a3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e6438189-ff16-4038-ba8d-3ee0972e45a3" (UID: "e6438189-ff16-4038-ba8d-3ee0972e45a3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.703392 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4f98d10-48e9-4d4f-abde-aef5db23a39e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f4f98d10-48e9-4d4f-abde-aef5db23a39e" (UID: "f4f98d10-48e9-4d4f-abde-aef5db23a39e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.707766 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad0e11b4-6dcf-437b-8798-1acdacc534ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad0e11b4-6dcf-437b-8798-1acdacc534ed" (UID: "ad0e11b4-6dcf-437b-8798-1acdacc534ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.707922 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629b86ef-b3c8-445c-aab0-c59fb2359f57-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "629b86ef-b3c8-445c-aab0-c59fb2359f57" (UID: "629b86ef-b3c8-445c-aab0-c59fb2359f57"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.708527 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13cbee2b-45bb-4d9e-8427-bb2574b8d284-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "13cbee2b-45bb-4d9e-8427-bb2574b8d284" (UID: "13cbee2b-45bb-4d9e-8427-bb2574b8d284"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.708664 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c2c10db-d15d-4fc4-b430-80b7cfb2c401" (UID: "3c2c10db-d15d-4fc4-b430-80b7cfb2c401"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.711447 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56f3777f-f5d6-4a05-855e-8a28b67a07eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "56f3777f-f5d6-4a05-855e-8a28b67a07eb" (UID: "56f3777f-f5d6-4a05-855e-8a28b67a07eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.713072 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4f98d10-48e9-4d4f-abde-aef5db23a39e-kube-api-access-pdk2m" (OuterVolumeSpecName: "kube-api-access-pdk2m") pod "f4f98d10-48e9-4d4f-abde-aef5db23a39e" (UID: "f4f98d10-48e9-4d4f-abde-aef5db23a39e"). InnerVolumeSpecName "kube-api-access-pdk2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.720934 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-kube-api-access-sr7sm" (OuterVolumeSpecName: "kube-api-access-sr7sm") pod "3c2c10db-d15d-4fc4-b430-80b7cfb2c401" (UID: "3c2c10db-d15d-4fc4-b430-80b7cfb2c401"). InnerVolumeSpecName "kube-api-access-sr7sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.721215 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad0e11b4-6dcf-437b-8798-1acdacc534ed-kube-api-access-rr7rj" (OuterVolumeSpecName: "kube-api-access-rr7rj") pod "ad0e11b4-6dcf-437b-8798-1acdacc534ed" (UID: "ad0e11b4-6dcf-437b-8798-1acdacc534ed"). InnerVolumeSpecName "kube-api-access-rr7rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.721468 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6438189-ff16-4038-ba8d-3ee0972e45a3-kube-api-access-fldl8" (OuterVolumeSpecName: "kube-api-access-fldl8") pod "e6438189-ff16-4038-ba8d-3ee0972e45a3" (UID: "e6438189-ff16-4038-ba8d-3ee0972e45a3"). InnerVolumeSpecName "kube-api-access-fldl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.722831 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/629b86ef-b3c8-445c-aab0-c59fb2359f57-kube-api-access-c2vtw" (OuterVolumeSpecName: "kube-api-access-c2vtw") pod "629b86ef-b3c8-445c-aab0-c59fb2359f57" (UID: "629b86ef-b3c8-445c-aab0-c59fb2359f57"). InnerVolumeSpecName "kube-api-access-c2vtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.723567 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56f3777f-f5d6-4a05-855e-8a28b67a07eb-kube-api-access-7hzmd" (OuterVolumeSpecName: "kube-api-access-7hzmd") pod "56f3777f-f5d6-4a05-855e-8a28b67a07eb" (UID: "56f3777f-f5d6-4a05-855e-8a28b67a07eb"). InnerVolumeSpecName "kube-api-access-7hzmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.742889 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13cbee2b-45bb-4d9e-8427-bb2574b8d284-kube-api-access-vnjw4" (OuterVolumeSpecName: "kube-api-access-vnjw4") pod "13cbee2b-45bb-4d9e-8427-bb2574b8d284" (UID: "13cbee2b-45bb-4d9e-8427-bb2574b8d284"). InnerVolumeSpecName "kube-api-access-vnjw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.804380 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk7zr\" (UniqueName: \"kubernetes.io/projected/6f22851f-5367-4fc7-92b4-6114a4d0449a-kube-api-access-xk7zr\") pod \"6f22851f-5367-4fc7-92b4-6114a4d0449a\" (UID: \"6f22851f-5367-4fc7-92b4-6114a4d0449a\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.804672 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f22851f-5367-4fc7-92b4-6114a4d0449a-operator-scripts\") pod \"6f22851f-5367-4fc7-92b4-6114a4d0449a\" (UID: \"6f22851f-5367-4fc7-92b4-6114a4d0449a\") " Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805565 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f22851f-5367-4fc7-92b4-6114a4d0449a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6f22851f-5367-4fc7-92b4-6114a4d0449a" (UID: "6f22851f-5367-4fc7-92b4-6114a4d0449a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805704 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13cbee2b-45bb-4d9e-8427-bb2574b8d284-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805731 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fldl8\" (UniqueName: \"kubernetes.io/projected/e6438189-ff16-4038-ba8d-3ee0972e45a3-kube-api-access-fldl8\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805746 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnjw4\" (UniqueName: \"kubernetes.io/projected/13cbee2b-45bb-4d9e-8427-bb2574b8d284-kube-api-access-vnjw4\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805758 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805769 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4f98d10-48e9-4d4f-abde-aef5db23a39e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805781 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hzmd\" (UniqueName: \"kubernetes.io/projected/56f3777f-f5d6-4a05-855e-8a28b67a07eb-kube-api-access-7hzmd\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805792 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad0e11b4-6dcf-437b-8798-1acdacc534ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805806 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/629b86ef-b3c8-445c-aab0-c59fb2359f57-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805817 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr7sm\" (UniqueName: \"kubernetes.io/projected/3c2c10db-d15d-4fc4-b430-80b7cfb2c401-kube-api-access-sr7sm\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805827 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2vtw\" (UniqueName: \"kubernetes.io/projected/629b86ef-b3c8-445c-aab0-c59fb2359f57-kube-api-access-c2vtw\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805839 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6438189-ff16-4038-ba8d-3ee0972e45a3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805853 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56f3777f-f5d6-4a05-855e-8a28b67a07eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805867 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdk2m\" (UniqueName: \"kubernetes.io/projected/f4f98d10-48e9-4d4f-abde-aef5db23a39e-kube-api-access-pdk2m\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.805881 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr7rj\" (UniqueName: \"kubernetes.io/projected/ad0e11b4-6dcf-437b-8798-1acdacc534ed-kube-api-access-rr7rj\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.809512 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-01f1-account-create-update-zd7hl" event={"ID":"f4f98d10-48e9-4d4f-abde-aef5db23a39e","Type":"ContainerDied","Data":"0821a0e84b1854e2a50f7695cf1595220b2d3e07317ad8c561a1ba6fb7c62d1c"} Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.809559 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0821a0e84b1854e2a50f7695cf1595220b2d3e07317ad8c561a1ba6fb7c62d1c" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.809657 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-01f1-account-create-update-zd7hl" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.810871 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f22851f-5367-4fc7-92b4-6114a4d0449a-kube-api-access-xk7zr" (OuterVolumeSpecName: "kube-api-access-xk7zr") pod "6f22851f-5367-4fc7-92b4-6114a4d0449a" (UID: "6f22851f-5367-4fc7-92b4-6114a4d0449a"). InnerVolumeSpecName "kube-api-access-xk7zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.823274 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" event={"ID":"629b86ef-b3c8-445c-aab0-c59fb2359f57","Type":"ContainerDied","Data":"9aeaf034113fc9f19026fd66f5e4e5884b17b2d5af026b7c7e7ac73923d88958"} Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.823334 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9aeaf034113fc9f19026fd66f5e4e5884b17b2d5af026b7c7e7ac73923d88958" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.823422 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0126-account-create-update-z7h68" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.825715 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c4jlk" event={"ID":"6f22851f-5367-4fc7-92b4-6114a4d0449a","Type":"ContainerDied","Data":"ed1068bf9ea9f65b955ac555a92eef385f4ec2f21629980060c99307fd2b58b9"} Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.825761 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed1068bf9ea9f65b955ac555a92eef385f4ec2f21629980060c99307fd2b58b9" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.825829 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c4jlk" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.835473 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8grd9" event={"ID":"ad0e11b4-6dcf-437b-8798-1acdacc534ed","Type":"ContainerDied","Data":"7f2c1c53aee0c8342e271ed52afd33b992e19106eeb559c786a2d2b993261957"} Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.835535 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f2c1c53aee0c8342e271ed52afd33b992e19106eeb559c786a2d2b993261957" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.835649 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8grd9" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.856271 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cfdd-account-create-update-tqjv4" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.857831 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cfdd-account-create-update-tqjv4" event={"ID":"e6438189-ff16-4038-ba8d-3ee0972e45a3","Type":"ContainerDied","Data":"efc61e3c6e70568469464b7484ce696bc18dcb14ec60fbd570128de34989a7ab"} Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.857874 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efc61e3c6e70568469464b7484ce696bc18dcb14ec60fbd570128de34989a7ab" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.870695 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.870689 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-5w9lz" event={"ID":"3c2c10db-d15d-4fc4-b430-80b7cfb2c401","Type":"ContainerDied","Data":"39ac039d13ac8d51a5fcfd8716a8777b47d72edfdd2c6e508d04a03053edcf47"} Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.870806 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39ac039d13ac8d51a5fcfd8716a8777b47d72edfdd2c6e508d04a03053edcf47" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.876675 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c8s82" event={"ID":"56f3777f-f5d6-4a05-855e-8a28b67a07eb","Type":"ContainerDied","Data":"7be2b29b530184976b6cd1da628b57dd1180642c1625978079fe5893c6d7ca0e"} Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.876719 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7be2b29b530184976b6cd1da628b57dd1180642c1625978079fe5893c6d7ca0e" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.876788 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c8s82" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.884846 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a366-account-create-update-zkvjq" event={"ID":"13cbee2b-45bb-4d9e-8427-bb2574b8d284","Type":"ContainerDied","Data":"7ebfa47bcec03ede20586aeab6282e41766775f1e5d6f90027f9dfd23e20f692"} Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.884895 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ebfa47bcec03ede20586aeab6282e41766775f1e5d6f90027f9dfd23e20f692" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.884982 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a366-account-create-update-zkvjq" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.908376 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f22851f-5367-4fc7-92b4-6114a4d0449a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:20 crc kubenswrapper[4954]: I1209 17:26:20.908420 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk7zr\" (UniqueName: \"kubernetes.io/projected/6f22851f-5367-4fc7-92b4-6114a4d0449a-kube-api-access-xk7zr\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:21 crc kubenswrapper[4954]: I1209 17:26:21.877894 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:26:21 crc kubenswrapper[4954]: I1209 17:26:21.935035 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-l6s5g" podUID="94a5a72d-c7dd-462f-969f-812555bc9ba3" containerName="ovn-controller" probeResult="failure" output=< Dec 09 17:26:21 crc kubenswrapper[4954]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 09 17:26:21 crc kubenswrapper[4954]: > Dec 09 17:26:22 crc kubenswrapper[4954]: I1209 17:26:22.624932 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:26:22 crc kubenswrapper[4954]: I1209 17:26:22.905564 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4vtb6" event={"ID":"9fc44917-47c6-450b-b393-bad021024fcd","Type":"ContainerStarted","Data":"8178577da1564319dcd83b00ff1b2453766fa8808cc19453311cd303d5135222"} Dec 09 17:26:22 crc kubenswrapper[4954]: I1209 17:26:22.909900 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f0da1b73-3249-449b-86f9-12206ce93aab","Type":"ContainerStarted","Data":"96f69de0abcceb205010d2c1050f29f7a03c6c5d48529e92e4467ff69964a002"} Dec 09 17:26:22 crc kubenswrapper[4954]: I1209 17:26:22.937080 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-4vtb6" podStartSLOduration=5.352460888 podStartE2EDuration="13.937054252s" podCreationTimestamp="2025-12-09 17:26:09 +0000 UTC" firstStartedPulling="2025-12-09 17:26:13.436192363 +0000 UTC m=+1769.824366183" lastFinishedPulling="2025-12-09 17:26:22.020785727 +0000 UTC m=+1778.408959547" observedRunningTime="2025-12-09 17:26:22.926227663 +0000 UTC m=+1779.314401483" watchObservedRunningTime="2025-12-09 17:26:22.937054252 +0000 UTC m=+1779.325228072" Dec 09 17:26:22 crc kubenswrapper[4954]: I1209 17:26:22.958047 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.587889599 podStartE2EDuration="1m5.958024108s" podCreationTimestamp="2025-12-09 17:25:17 +0000 UTC" firstStartedPulling="2025-12-09 17:25:33.741062509 +0000 UTC m=+1730.129236359" lastFinishedPulling="2025-12-09 17:26:22.111197048 +0000 UTC m=+1778.499370868" observedRunningTime="2025-12-09 17:26:22.953420794 +0000 UTC m=+1779.341594614" watchObservedRunningTime="2025-12-09 17:26:22.958024108 +0000 UTC m=+1779.346197928" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.050959 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.123113 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.124096 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.151660 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-n5hmc"] Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.152101 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" podUID="a767b973-ed53-4675-93a8-db722496f8d7" containerName="dnsmasq-dns" containerID="cri-o://9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f" gracePeriod=10 Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.242672 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn"] Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243431 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6438189-ff16-4038-ba8d-3ee0972e45a3" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243455 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6438189-ff16-4038-ba8d-3ee0972e45a3" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243490 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb473662-281d-46be-996e-02370393136b" containerName="console" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243498 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb473662-281d-46be-996e-02370393136b" containerName="console" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243510 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4f98d10-48e9-4d4f-abde-aef5db23a39e" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243516 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4f98d10-48e9-4d4f-abde-aef5db23a39e" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243526 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d66ceb0-8567-43c3-bbda-46693b51be82" containerName="init" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243532 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d66ceb0-8567-43c3-bbda-46693b51be82" containerName="init" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243546 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629b86ef-b3c8-445c-aab0-c59fb2359f57" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243553 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="629b86ef-b3c8-445c-aab0-c59fb2359f57" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243562 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f3777f-f5d6-4a05-855e-8a28b67a07eb" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243572 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f3777f-f5d6-4a05-855e-8a28b67a07eb" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243586 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d66ceb0-8567-43c3-bbda-46693b51be82" containerName="dnsmasq-dns" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243608 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d66ceb0-8567-43c3-bbda-46693b51be82" containerName="dnsmasq-dns" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243623 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13cbee2b-45bb-4d9e-8427-bb2574b8d284" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243629 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="13cbee2b-45bb-4d9e-8427-bb2574b8d284" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243640 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f22851f-5367-4fc7-92b4-6114a4d0449a" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243647 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f22851f-5367-4fc7-92b4-6114a4d0449a" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243657 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0e11b4-6dcf-437b-8798-1acdacc534ed" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243663 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0e11b4-6dcf-437b-8798-1acdacc534ed" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: E1209 17:26:23.243676 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c2c10db-d15d-4fc4-b430-80b7cfb2c401" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243684 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c2c10db-d15d-4fc4-b430-80b7cfb2c401" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243898 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4f98d10-48e9-4d4f-abde-aef5db23a39e" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243912 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c2c10db-d15d-4fc4-b430-80b7cfb2c401" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243926 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb473662-281d-46be-996e-02370393136b" containerName="console" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243936 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="56f3777f-f5d6-4a05-855e-8a28b67a07eb" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243948 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d66ceb0-8567-43c3-bbda-46693b51be82" containerName="dnsmasq-dns" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243958 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f22851f-5367-4fc7-92b4-6114a4d0449a" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243969 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0e11b4-6dcf-437b-8798-1acdacc534ed" containerName="mariadb-database-create" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243979 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="13cbee2b-45bb-4d9e-8427-bb2574b8d284" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243986 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6438189-ff16-4038-ba8d-3ee0972e45a3" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.243996 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="629b86ef-b3c8-445c-aab0-c59fb2359f57" containerName="mariadb-account-create-update" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.244863 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.249856 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn"] Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.308352 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9hpt\" (UniqueName: \"kubernetes.io/projected/16ad1f90-16e6-4bf8-bdee-509182d1e152-kube-api-access-g9hpt\") pod \"mysqld-exporter-openstack-cell1-db-create-gzgnn\" (UID: \"16ad1f90-16e6-4bf8-bdee-509182d1e152\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.308441 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ad1f90-16e6-4bf8-bdee-509182d1e152-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-gzgnn\" (UID: \"16ad1f90-16e6-4bf8-bdee-509182d1e152\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.410789 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9hpt\" (UniqueName: \"kubernetes.io/projected/16ad1f90-16e6-4bf8-bdee-509182d1e152-kube-api-access-g9hpt\") pod \"mysqld-exporter-openstack-cell1-db-create-gzgnn\" (UID: \"16ad1f90-16e6-4bf8-bdee-509182d1e152\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.410851 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ad1f90-16e6-4bf8-bdee-509182d1e152-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-gzgnn\" (UID: \"16ad1f90-16e6-4bf8-bdee-509182d1e152\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.411944 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ad1f90-16e6-4bf8-bdee-509182d1e152-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-gzgnn\" (UID: \"16ad1f90-16e6-4bf8-bdee-509182d1e152\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.463388 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9hpt\" (UniqueName: \"kubernetes.io/projected/16ad1f90-16e6-4bf8-bdee-509182d1e152-kube-api-access-g9hpt\") pod \"mysqld-exporter-openstack-cell1-db-create-gzgnn\" (UID: \"16ad1f90-16e6-4bf8-bdee-509182d1e152\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.470900 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-2e7f-account-create-update-hh2nj"] Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.472410 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.478818 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.513011 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v25r\" (UniqueName: \"kubernetes.io/projected/3df17d5e-f6cc-46af-b015-8640baaa8d45-kube-api-access-7v25r\") pod \"mysqld-exporter-2e7f-account-create-update-hh2nj\" (UID: \"3df17d5e-f6cc-46af-b015-8640baaa8d45\") " pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.513501 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3df17d5e-f6cc-46af-b015-8640baaa8d45-operator-scripts\") pod \"mysqld-exporter-2e7f-account-create-update-hh2nj\" (UID: \"3df17d5e-f6cc-46af-b015-8640baaa8d45\") " pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.526686 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-2e7f-account-create-update-hh2nj"] Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.616421 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v25r\" (UniqueName: \"kubernetes.io/projected/3df17d5e-f6cc-46af-b015-8640baaa8d45-kube-api-access-7v25r\") pod \"mysqld-exporter-2e7f-account-create-update-hh2nj\" (UID: \"3df17d5e-f6cc-46af-b015-8640baaa8d45\") " pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.617239 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3df17d5e-f6cc-46af-b015-8640baaa8d45-operator-scripts\") pod \"mysqld-exporter-2e7f-account-create-update-hh2nj\" (UID: \"3df17d5e-f6cc-46af-b015-8640baaa8d45\") " pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.618070 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3df17d5e-f6cc-46af-b015-8640baaa8d45-operator-scripts\") pod \"mysqld-exporter-2e7f-account-create-update-hh2nj\" (UID: \"3df17d5e-f6cc-46af-b015-8640baaa8d45\") " pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.631766 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.637669 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v25r\" (UniqueName: \"kubernetes.io/projected/3df17d5e-f6cc-46af-b015-8640baaa8d45-kube-api-access-7v25r\") pod \"mysqld-exporter-2e7f-account-create-update-hh2nj\" (UID: \"3df17d5e-f6cc-46af-b015-8640baaa8d45\") " pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.802118 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.822011 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-sb\") pod \"a767b973-ed53-4675-93a8-db722496f8d7\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.822066 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-nb\") pod \"a767b973-ed53-4675-93a8-db722496f8d7\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.822132 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-config\") pod \"a767b973-ed53-4675-93a8-db722496f8d7\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.822190 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bspsr\" (UniqueName: \"kubernetes.io/projected/a767b973-ed53-4675-93a8-db722496f8d7-kube-api-access-bspsr\") pod \"a767b973-ed53-4675-93a8-db722496f8d7\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.822276 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-dns-svc\") pod \"a767b973-ed53-4675-93a8-db722496f8d7\" (UID: \"a767b973-ed53-4675-93a8-db722496f8d7\") " Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.860308 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a767b973-ed53-4675-93a8-db722496f8d7-kube-api-access-bspsr" (OuterVolumeSpecName: "kube-api-access-bspsr") pod "a767b973-ed53-4675-93a8-db722496f8d7" (UID: "a767b973-ed53-4675-93a8-db722496f8d7"). InnerVolumeSpecName "kube-api-access-bspsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.874717 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a767b973-ed53-4675-93a8-db722496f8d7" (UID: "a767b973-ed53-4675-93a8-db722496f8d7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.891357 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.924334 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.924374 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bspsr\" (UniqueName: \"kubernetes.io/projected/a767b973-ed53-4675-93a8-db722496f8d7-kube-api-access-bspsr\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.930945 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-config" (OuterVolumeSpecName: "config") pod "a767b973-ed53-4675-93a8-db722496f8d7" (UID: "a767b973-ed53-4675-93a8-db722496f8d7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.950839 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a767b973-ed53-4675-93a8-db722496f8d7" (UID: "a767b973-ed53-4675-93a8-db722496f8d7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.954366 4954 generic.go:334] "Generic (PLEG): container finished" podID="3645ed14-3d97-496e-a683-2e56ea33fec9" containerID="e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6" exitCode=0 Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.954457 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3645ed14-3d97-496e-a683-2e56ea33fec9","Type":"ContainerDied","Data":"e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6"} Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.956218 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a767b973-ed53-4675-93a8-db722496f8d7" (UID: "a767b973-ed53-4675-93a8-db722496f8d7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.963701 4954 generic.go:334] "Generic (PLEG): container finished" podID="a767b973-ed53-4675-93a8-db722496f8d7" containerID="9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f" exitCode=0 Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.964288 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" event={"ID":"a767b973-ed53-4675-93a8-db722496f8d7","Type":"ContainerDied","Data":"9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f"} Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.964335 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.964379 4954 scope.go:117] "RemoveContainer" containerID="9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f" Dec 09 17:26:23 crc kubenswrapper[4954]: I1209 17:26:23.964358 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-n5hmc" event={"ID":"a767b973-ed53-4675-93a8-db722496f8d7","Type":"ContainerDied","Data":"fef072276e33a16173984619b7e235b22f7313fda5e2f16bebcaf81bd34824c3"} Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.027958 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.028013 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.028027 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a767b973-ed53-4675-93a8-db722496f8d7-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.042094 4954 scope.go:117] "RemoveContainer" containerID="340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308" Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.053142 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-n5hmc"] Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.074052 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-n5hmc"] Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.126921 4954 scope.go:117] "RemoveContainer" containerID="9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f" Dec 09 17:26:24 crc kubenswrapper[4954]: E1209 17:26:24.128339 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f\": container with ID starting with 9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f not found: ID does not exist" containerID="9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f" Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.128392 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f"} err="failed to get container status \"9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f\": rpc error: code = NotFound desc = could not find container \"9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f\": container with ID starting with 9f3902a0e768dfcbe12feca6a2a2338c1968ab717c749af191dd82136deb782f not found: ID does not exist" Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.128423 4954 scope.go:117] "RemoveContainer" containerID="340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308" Dec 09 17:26:24 crc kubenswrapper[4954]: E1209 17:26:24.131534 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308\": container with ID starting with 340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308 not found: ID does not exist" containerID="340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308" Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.131677 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308"} err="failed to get container status \"340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308\": rpc error: code = NotFound desc = could not find container \"340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308\": container with ID starting with 340c4868b85faf47f3219ee2dc4542d719fad9d23e955fe0d08119d11a6eb308 not found: ID does not exist" Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.153611 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a767b973-ed53-4675-93a8-db722496f8d7" path="/var/lib/kubelet/pods/a767b973-ed53-4675-93a8-db722496f8d7/volumes" Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.274968 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn"] Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.519814 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-2e7f-account-create-update-hh2nj"] Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.798704 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.949342 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:24 crc kubenswrapper[4954]: E1209 17:26:24.949600 4954 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 09 17:26:24 crc kubenswrapper[4954]: E1209 17:26:24.949635 4954 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 09 17:26:24 crc kubenswrapper[4954]: E1209 17:26:24.949709 4954 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift podName:69b248e7-6cb4-4805-a01f-537fe58aa42d nodeName:}" failed. No retries permitted until 2025-12-09 17:26:40.949684328 +0000 UTC m=+1797.337858148 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift") pod "swift-storage-0" (UID: "69b248e7-6cb4-4805-a01f-537fe58aa42d") : configmap "swift-ring-files" not found Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.974134 4954 generic.go:334] "Generic (PLEG): container finished" podID="16ad1f90-16e6-4bf8-bdee-509182d1e152" containerID="9d55d5d0a551d4efa78f31acb42645047cd58b12a48c7b9fd0bcd9b737fcf22c" exitCode=0 Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.974201 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" event={"ID":"16ad1f90-16e6-4bf8-bdee-509182d1e152","Type":"ContainerDied","Data":"9d55d5d0a551d4efa78f31acb42645047cd58b12a48c7b9fd0bcd9b737fcf22c"} Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.974233 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" event={"ID":"16ad1f90-16e6-4bf8-bdee-509182d1e152","Type":"ContainerStarted","Data":"a3b129ff32cfcf490eeb20bbd83e79db57a097fb92becef76fae0f62d3a55a06"} Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.976828 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" event={"ID":"3df17d5e-f6cc-46af-b015-8640baaa8d45","Type":"ContainerStarted","Data":"c22fbc0e99b17e5023a9950dcfa5eda459b4dbbfa399207e2360dc07acd4753e"} Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.976863 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" event={"ID":"3df17d5e-f6cc-46af-b015-8640baaa8d45","Type":"ContainerStarted","Data":"92c4d067a4ad93353c7e4d0cc7c4c6baa8ee8eaedbfa2fda567773ef9ef12277"} Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.980988 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3645ed14-3d97-496e-a683-2e56ea33fec9","Type":"ContainerStarted","Data":"e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366"} Dec 09 17:26:24 crc kubenswrapper[4954]: I1209 17:26:24.982561 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 09 17:26:25 crc kubenswrapper[4954]: I1209 17:26:25.081344 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=75.081317789 podStartE2EDuration="1m15.081317789s" podCreationTimestamp="2025-12-09 17:25:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:25.078229442 +0000 UTC m=+1781.466403272" watchObservedRunningTime="2025-12-09 17:26:25.081317789 +0000 UTC m=+1781.469491609" Dec 09 17:26:25 crc kubenswrapper[4954]: I1209 17:26:25.123783 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" podStartSLOduration=2.123755197 podStartE2EDuration="2.123755197s" podCreationTimestamp="2025-12-09 17:26:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:25.11809653 +0000 UTC m=+1781.506270340" watchObservedRunningTime="2025-12-09 17:26:25.123755197 +0000 UTC m=+1781.511929017" Dec 09 17:26:25 crc kubenswrapper[4954]: I1209 17:26:25.991716 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-g4kjd"] Dec 09 17:26:25 crc kubenswrapper[4954]: E1209 17:26:25.992851 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a767b973-ed53-4675-93a8-db722496f8d7" containerName="init" Dec 09 17:26:25 crc kubenswrapper[4954]: I1209 17:26:25.992870 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a767b973-ed53-4675-93a8-db722496f8d7" containerName="init" Dec 09 17:26:25 crc kubenswrapper[4954]: E1209 17:26:25.992894 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a767b973-ed53-4675-93a8-db722496f8d7" containerName="dnsmasq-dns" Dec 09 17:26:25 crc kubenswrapper[4954]: I1209 17:26:25.992903 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a767b973-ed53-4675-93a8-db722496f8d7" containerName="dnsmasq-dns" Dec 09 17:26:25 crc kubenswrapper[4954]: I1209 17:26:25.993180 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a767b973-ed53-4675-93a8-db722496f8d7" containerName="dnsmasq-dns" Dec 09 17:26:25 crc kubenswrapper[4954]: I1209 17:26:25.994239 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:25 crc kubenswrapper[4954]: I1209 17:26:25.998785 4954 generic.go:334] "Generic (PLEG): container finished" podID="3df17d5e-f6cc-46af-b015-8640baaa8d45" containerID="c22fbc0e99b17e5023a9950dcfa5eda459b4dbbfa399207e2360dc07acd4753e" exitCode=0 Dec 09 17:26:25 crc kubenswrapper[4954]: I1209 17:26:25.999088 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" event={"ID":"3df17d5e-f6cc-46af-b015-8640baaa8d45","Type":"ContainerDied","Data":"c22fbc0e99b17e5023a9950dcfa5eda459b4dbbfa399207e2360dc07acd4753e"} Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.000754 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.010585 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mprmf" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.028887 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g4kjd"] Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.104256 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-db-sync-config-data\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.104357 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-combined-ca-bundle\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.104436 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wz8q\" (UniqueName: \"kubernetes.io/projected/8640fe28-2072-49f7-8c8d-35a2ea5432e0-kube-api-access-7wz8q\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.104739 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-config-data\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.206706 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-config-data\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.206951 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-db-sync-config-data\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.207034 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-combined-ca-bundle\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.207163 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wz8q\" (UniqueName: \"kubernetes.io/projected/8640fe28-2072-49f7-8c8d-35a2ea5432e0-kube-api-access-7wz8q\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.213835 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-db-sync-config-data\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.214397 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-combined-ca-bundle\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.222300 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-config-data\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.253214 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wz8q\" (UniqueName: \"kubernetes.io/projected/8640fe28-2072-49f7-8c8d-35a2ea5432e0-kube-api-access-7wz8q\") pod \"glance-db-sync-g4kjd\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.319406 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.510230 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.615960 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ad1f90-16e6-4bf8-bdee-509182d1e152-operator-scripts\") pod \"16ad1f90-16e6-4bf8-bdee-509182d1e152\" (UID: \"16ad1f90-16e6-4bf8-bdee-509182d1e152\") " Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.616097 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9hpt\" (UniqueName: \"kubernetes.io/projected/16ad1f90-16e6-4bf8-bdee-509182d1e152-kube-api-access-g9hpt\") pod \"16ad1f90-16e6-4bf8-bdee-509182d1e152\" (UID: \"16ad1f90-16e6-4bf8-bdee-509182d1e152\") " Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.618488 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16ad1f90-16e6-4bf8-bdee-509182d1e152-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16ad1f90-16e6-4bf8-bdee-509182d1e152" (UID: "16ad1f90-16e6-4bf8-bdee-509182d1e152"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.628465 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16ad1f90-16e6-4bf8-bdee-509182d1e152-kube-api-access-g9hpt" (OuterVolumeSpecName: "kube-api-access-g9hpt") pod "16ad1f90-16e6-4bf8-bdee-509182d1e152" (UID: "16ad1f90-16e6-4bf8-bdee-509182d1e152"). InnerVolumeSpecName "kube-api-access-g9hpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.719241 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ad1f90-16e6-4bf8-bdee-509182d1e152-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.719298 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9hpt\" (UniqueName: \"kubernetes.io/projected/16ad1f90-16e6-4bf8-bdee-509182d1e152-kube-api-access-g9hpt\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.772045 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g4kjd"] Dec 09 17:26:26 crc kubenswrapper[4954]: W1209 17:26:26.773452 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8640fe28_2072_49f7_8c8d_35a2ea5432e0.slice/crio-d11de16b9e4b72b97a53e9c762db429bdb3bba6c9c86c62a0c31a413954fc1a4 WatchSource:0}: Error finding container d11de16b9e4b72b97a53e9c762db429bdb3bba6c9c86c62a0c31a413954fc1a4: Status 404 returned error can't find the container with id d11de16b9e4b72b97a53e9c762db429bdb3bba6c9c86c62a0c31a413954fc1a4 Dec 09 17:26:26 crc kubenswrapper[4954]: I1209 17:26:26.931446 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-l6s5g" podUID="94a5a72d-c7dd-462f-969f-812555bc9ba3" containerName="ovn-controller" probeResult="failure" output=< Dec 09 17:26:26 crc kubenswrapper[4954]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 09 17:26:26 crc kubenswrapper[4954]: > Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.010760 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g4kjd" event={"ID":"8640fe28-2072-49f7-8c8d-35a2ea5432e0","Type":"ContainerStarted","Data":"d11de16b9e4b72b97a53e9c762db429bdb3bba6c9c86c62a0c31a413954fc1a4"} Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.012783 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.012782 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn" event={"ID":"16ad1f90-16e6-4bf8-bdee-509182d1e152","Type":"ContainerDied","Data":"a3b129ff32cfcf490eeb20bbd83e79db57a097fb92becef76fae0f62d3a55a06"} Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.012855 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3b129ff32cfcf490eeb20bbd83e79db57a097fb92becef76fae0f62d3a55a06" Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.564973 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.676371 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-ggbdh" Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.748984 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v25r\" (UniqueName: \"kubernetes.io/projected/3df17d5e-f6cc-46af-b015-8640baaa8d45-kube-api-access-7v25r\") pod \"3df17d5e-f6cc-46af-b015-8640baaa8d45\" (UID: \"3df17d5e-f6cc-46af-b015-8640baaa8d45\") " Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.749547 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3df17d5e-f6cc-46af-b015-8640baaa8d45-operator-scripts\") pod \"3df17d5e-f6cc-46af-b015-8640baaa8d45\" (UID: \"3df17d5e-f6cc-46af-b015-8640baaa8d45\") " Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.751263 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3df17d5e-f6cc-46af-b015-8640baaa8d45-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3df17d5e-f6cc-46af-b015-8640baaa8d45" (UID: "3df17d5e-f6cc-46af-b015-8640baaa8d45"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.800786 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3df17d5e-f6cc-46af-b015-8640baaa8d45-kube-api-access-7v25r" (OuterVolumeSpecName: "kube-api-access-7v25r") pod "3df17d5e-f6cc-46af-b015-8640baaa8d45" (UID: "3df17d5e-f6cc-46af-b015-8640baaa8d45"). InnerVolumeSpecName "kube-api-access-7v25r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.852136 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v25r\" (UniqueName: \"kubernetes.io/projected/3df17d5e-f6cc-46af-b015-8640baaa8d45-kube-api-access-7v25r\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:27 crc kubenswrapper[4954]: I1209 17:26:27.852177 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3df17d5e-f6cc-46af-b015-8640baaa8d45-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.032415 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-l6s5g-config-854zf"] Dec 09 17:26:28 crc kubenswrapper[4954]: E1209 17:26:28.033018 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df17d5e-f6cc-46af-b015-8640baaa8d45" containerName="mariadb-account-create-update" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.033033 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df17d5e-f6cc-46af-b015-8640baaa8d45" containerName="mariadb-account-create-update" Dec 09 17:26:28 crc kubenswrapper[4954]: E1209 17:26:28.033054 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ad1f90-16e6-4bf8-bdee-509182d1e152" containerName="mariadb-database-create" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.033064 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ad1f90-16e6-4bf8-bdee-509182d1e152" containerName="mariadb-database-create" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.033295 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="16ad1f90-16e6-4bf8-bdee-509182d1e152" containerName="mariadb-database-create" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.033313 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df17d5e-f6cc-46af-b015-8640baaa8d45" containerName="mariadb-account-create-update" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.034200 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.036489 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" event={"ID":"3df17d5e-f6cc-46af-b015-8640baaa8d45","Type":"ContainerDied","Data":"92c4d067a4ad93353c7e4d0cc7c4c6baa8ee8eaedbfa2fda567773ef9ef12277"} Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.036541 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92c4d067a4ad93353c7e4d0cc7c4c6baa8ee8eaedbfa2fda567773ef9ef12277" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.036630 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-2e7f-account-create-update-hh2nj" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.038954 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.092052 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l6s5g-config-854zf"] Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.160249 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-log-ovn\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.160363 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s89fv\" (UniqueName: \"kubernetes.io/projected/91cac29e-62c0-4887-824d-bd65e665e694-kube-api-access-s89fv\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.160481 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-additional-scripts\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.160852 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.161495 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-scripts\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.161846 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run-ovn\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.264446 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-scripts\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.264859 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run-ovn\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.264900 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-log-ovn\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.264923 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s89fv\" (UniqueName: \"kubernetes.io/projected/91cac29e-62c0-4887-824d-bd65e665e694-kube-api-access-s89fv\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.264952 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-additional-scripts\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.265000 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.265354 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.265419 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run-ovn\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.265454 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-log-ovn\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.266450 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-additional-scripts\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.266964 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-scripts\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.289681 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s89fv\" (UniqueName: \"kubernetes.io/projected/91cac29e-62c0-4887-824d-bd65e665e694-kube-api-access-s89fv\") pod \"ovn-controller-l6s5g-config-854zf\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:28 crc kubenswrapper[4954]: I1209 17:26:28.402297 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:29 crc kubenswrapper[4954]: I1209 17:26:29.063888 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l6s5g-config-854zf"] Dec 09 17:26:30 crc kubenswrapper[4954]: I1209 17:26:30.059584 4954 generic.go:334] "Generic (PLEG): container finished" podID="91cac29e-62c0-4887-824d-bd65e665e694" containerID="7288584ac020f9f057251034402d4eeca1bba70fe00639b87d8c6bac614ed61e" exitCode=0 Dec 09 17:26:30 crc kubenswrapper[4954]: I1209 17:26:30.059739 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l6s5g-config-854zf" event={"ID":"91cac29e-62c0-4887-824d-bd65e665e694","Type":"ContainerDied","Data":"7288584ac020f9f057251034402d4eeca1bba70fe00639b87d8c6bac614ed61e"} Dec 09 17:26:30 crc kubenswrapper[4954]: I1209 17:26:30.059940 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l6s5g-config-854zf" event={"ID":"91cac29e-62c0-4887-824d-bd65e665e694","Type":"ContainerStarted","Data":"93142e1c2424aab813cbac5995a958abb5630939c8cae5ec6c97b86c79fcd75b"} Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.483287 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.646841 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-log-ovn\") pod \"91cac29e-62c0-4887-824d-bd65e665e694\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.646949 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-additional-scripts\") pod \"91cac29e-62c0-4887-824d-bd65e665e694\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.646995 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "91cac29e-62c0-4887-824d-bd65e665e694" (UID: "91cac29e-62c0-4887-824d-bd65e665e694"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.647068 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s89fv\" (UniqueName: \"kubernetes.io/projected/91cac29e-62c0-4887-824d-bd65e665e694-kube-api-access-s89fv\") pod \"91cac29e-62c0-4887-824d-bd65e665e694\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.647173 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run\") pod \"91cac29e-62c0-4887-824d-bd65e665e694\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.647271 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-scripts\") pod \"91cac29e-62c0-4887-824d-bd65e665e694\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.647299 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run-ovn\") pod \"91cac29e-62c0-4887-824d-bd65e665e694\" (UID: \"91cac29e-62c0-4887-824d-bd65e665e694\") " Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.647403 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run" (OuterVolumeSpecName: "var-run") pod "91cac29e-62c0-4887-824d-bd65e665e694" (UID: "91cac29e-62c0-4887-824d-bd65e665e694"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.647537 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "91cac29e-62c0-4887-824d-bd65e665e694" (UID: "91cac29e-62c0-4887-824d-bd65e665e694"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.648716 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-scripts" (OuterVolumeSpecName: "scripts") pod "91cac29e-62c0-4887-824d-bd65e665e694" (UID: "91cac29e-62c0-4887-824d-bd65e665e694"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.649403 4954 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.649431 4954 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.649435 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "91cac29e-62c0-4887-824d-bd65e665e694" (UID: "91cac29e-62c0-4887-824d-bd65e665e694"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.649443 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.649471 4954 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91cac29e-62c0-4887-824d-bd65e665e694-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.654242 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91cac29e-62c0-4887-824d-bd65e665e694-kube-api-access-s89fv" (OuterVolumeSpecName: "kube-api-access-s89fv") pod "91cac29e-62c0-4887-824d-bd65e665e694" (UID: "91cac29e-62c0-4887-824d-bd65e665e694"). InnerVolumeSpecName "kube-api-access-s89fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.751894 4954 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/91cac29e-62c0-4887-824d-bd65e665e694-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.751935 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s89fv\" (UniqueName: \"kubernetes.io/projected/91cac29e-62c0-4887-824d-bd65e665e694-kube-api-access-s89fv\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:31 crc kubenswrapper[4954]: I1209 17:26:31.938768 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-l6s5g" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.094057 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l6s5g-config-854zf" event={"ID":"91cac29e-62c0-4887-824d-bd65e665e694","Type":"ContainerDied","Data":"93142e1c2424aab813cbac5995a958abb5630939c8cae5ec6c97b86c79fcd75b"} Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.094100 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l6s5g-config-854zf" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.094122 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93142e1c2424aab813cbac5995a958abb5630939c8cae5ec6c97b86c79fcd75b" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.627496 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-l6s5g-config-854zf"] Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.646444 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-l6s5g-config-854zf"] Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.734803 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-l6s5g-config-gn9nb"] Dec 09 17:26:32 crc kubenswrapper[4954]: E1209 17:26:32.735465 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91cac29e-62c0-4887-824d-bd65e665e694" containerName="ovn-config" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.735493 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="91cac29e-62c0-4887-824d-bd65e665e694" containerName="ovn-config" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.735781 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="91cac29e-62c0-4887-824d-bd65e665e694" containerName="ovn-config" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.736824 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.739308 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.751426 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l6s5g-config-gn9nb"] Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.879873 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.879979 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run-ovn\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.880014 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz8pm\" (UniqueName: \"kubernetes.io/projected/559ebfa2-b84e-4b18-901d-314bdddd8fcc-kube-api-access-rz8pm\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.880090 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-log-ovn\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.880141 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-scripts\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.880183 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-additional-scripts\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.982754 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-log-ovn\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.982877 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-scripts\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.982947 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-additional-scripts\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.983054 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.983143 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run-ovn\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.983177 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz8pm\" (UniqueName: \"kubernetes.io/projected/559ebfa2-b84e-4b18-901d-314bdddd8fcc-kube-api-access-rz8pm\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.985694 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-log-ovn\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.988489 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-scripts\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.989078 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-additional-scripts\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.989160 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:32 crc kubenswrapper[4954]: I1209 17:26:32.989209 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run-ovn\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.024977 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz8pm\" (UniqueName: \"kubernetes.io/projected/559ebfa2-b84e-4b18-901d-314bdddd8fcc-kube-api-access-rz8pm\") pod \"ovn-controller-l6s5g-config-gn9nb\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.080629 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.117488 4954 generic.go:334] "Generic (PLEG): container finished" podID="9fc44917-47c6-450b-b393-bad021024fcd" containerID="8178577da1564319dcd83b00ff1b2453766fa8808cc19453311cd303d5135222" exitCode=0 Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.117540 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4vtb6" event={"ID":"9fc44917-47c6-450b-b393-bad021024fcd","Type":"ContainerDied","Data":"8178577da1564319dcd83b00ff1b2453766fa8808cc19453311cd303d5135222"} Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.605799 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l6s5g-config-gn9nb"] Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.768789 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.770181 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.774782 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.787585 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.916555 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " pod="openstack/mysqld-exporter-0" Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.917085 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rwtc\" (UniqueName: \"kubernetes.io/projected/ce89f21f-56e2-4107-ae7a-cb38603a2e01-kube-api-access-8rwtc\") pod \"mysqld-exporter-0\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " pod="openstack/mysqld-exporter-0" Dec 09 17:26:33 crc kubenswrapper[4954]: I1209 17:26:33.917206 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-config-data\") pod \"mysqld-exporter-0\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " pod="openstack/mysqld-exporter-0" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.019102 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rwtc\" (UniqueName: \"kubernetes.io/projected/ce89f21f-56e2-4107-ae7a-cb38603a2e01-kube-api-access-8rwtc\") pod \"mysqld-exporter-0\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " pod="openstack/mysqld-exporter-0" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.019446 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-config-data\") pod \"mysqld-exporter-0\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " pod="openstack/mysqld-exporter-0" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.019516 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " pod="openstack/mysqld-exporter-0" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.038292 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-config-data\") pod \"mysqld-exporter-0\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " pod="openstack/mysqld-exporter-0" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.040836 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rwtc\" (UniqueName: \"kubernetes.io/projected/ce89f21f-56e2-4107-ae7a-cb38603a2e01-kube-api-access-8rwtc\") pod \"mysqld-exporter-0\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " pod="openstack/mysqld-exporter-0" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.059096 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " pod="openstack/mysqld-exporter-0" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.096124 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.144323 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91cac29e-62c0-4887-824d-bd65e665e694" path="/var/lib/kubelet/pods/91cac29e-62c0-4887-824d-bd65e665e694/volumes" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.154424 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l6s5g-config-gn9nb" event={"ID":"559ebfa2-b84e-4b18-901d-314bdddd8fcc","Type":"ContainerStarted","Data":"35b1cdfbacb3a5a691cd62bbe33d3dc479c80cc894a88016c2cf307a430ecd18"} Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.154500 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l6s5g-config-gn9nb" event={"ID":"559ebfa2-b84e-4b18-901d-314bdddd8fcc","Type":"ContainerStarted","Data":"ea39cf07b2eb95c699b2f29b4dd245d16adcebea30009013e7cd484c04892421"} Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.245011 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-l6s5g-config-gn9nb" podStartSLOduration=2.244982001 podStartE2EDuration="2.244982001s" podCreationTimestamp="2025-12-09 17:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:34.206899258 +0000 UTC m=+1790.595073078" watchObservedRunningTime="2025-12-09 17:26:34.244982001 +0000 UTC m=+1790.633155821" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.736076 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.795475 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.799395 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.804006 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.850655 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-scripts\") pod \"9fc44917-47c6-450b-b393-bad021024fcd\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.850918 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fc44917-47c6-450b-b393-bad021024fcd-etc-swift\") pod \"9fc44917-47c6-450b-b393-bad021024fcd\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.850966 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-swiftconf\") pod \"9fc44917-47c6-450b-b393-bad021024fcd\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.851016 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-ring-data-devices\") pod \"9fc44917-47c6-450b-b393-bad021024fcd\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.851075 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-496j9\" (UniqueName: \"kubernetes.io/projected/9fc44917-47c6-450b-b393-bad021024fcd-kube-api-access-496j9\") pod \"9fc44917-47c6-450b-b393-bad021024fcd\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.851149 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-combined-ca-bundle\") pod \"9fc44917-47c6-450b-b393-bad021024fcd\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.851196 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-dispersionconf\") pod \"9fc44917-47c6-450b-b393-bad021024fcd\" (UID: \"9fc44917-47c6-450b-b393-bad021024fcd\") " Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.852509 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9fc44917-47c6-450b-b393-bad021024fcd" (UID: "9fc44917-47c6-450b-b393-bad021024fcd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.853412 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fc44917-47c6-450b-b393-bad021024fcd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9fc44917-47c6-450b-b393-bad021024fcd" (UID: "9fc44917-47c6-450b-b393-bad021024fcd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.856961 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fc44917-47c6-450b-b393-bad021024fcd-kube-api-access-496j9" (OuterVolumeSpecName: "kube-api-access-496j9") pod "9fc44917-47c6-450b-b393-bad021024fcd" (UID: "9fc44917-47c6-450b-b393-bad021024fcd"). InnerVolumeSpecName "kube-api-access-496j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.861254 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9fc44917-47c6-450b-b393-bad021024fcd" (UID: "9fc44917-47c6-450b-b393-bad021024fcd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.883510 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-scripts" (OuterVolumeSpecName: "scripts") pod "9fc44917-47c6-450b-b393-bad021024fcd" (UID: "9fc44917-47c6-450b-b393-bad021024fcd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.884875 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9fc44917-47c6-450b-b393-bad021024fcd" (UID: "9fc44917-47c6-450b-b393-bad021024fcd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.905098 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fc44917-47c6-450b-b393-bad021024fcd" (UID: "9fc44917-47c6-450b-b393-bad021024fcd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.954053 4954 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9fc44917-47c6-450b-b393-bad021024fcd-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.954096 4954 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.954110 4954 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.954123 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-496j9\" (UniqueName: \"kubernetes.io/projected/9fc44917-47c6-450b-b393-bad021024fcd-kube-api-access-496j9\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.954137 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.954149 4954 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9fc44917-47c6-450b-b393-bad021024fcd-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:34 crc kubenswrapper[4954]: I1209 17:26:34.954291 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fc44917-47c6-450b-b393-bad021024fcd-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:35 crc kubenswrapper[4954]: I1209 17:26:35.184514 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"ce89f21f-56e2-4107-ae7a-cb38603a2e01","Type":"ContainerStarted","Data":"563b1afc7a6a0927b41c3a37f441088d67e2657c116c6f1f174dfcc05155e0c0"} Dec 09 17:26:35 crc kubenswrapper[4954]: I1209 17:26:35.225638 4954 generic.go:334] "Generic (PLEG): container finished" podID="559ebfa2-b84e-4b18-901d-314bdddd8fcc" containerID="35b1cdfbacb3a5a691cd62bbe33d3dc479c80cc894a88016c2cf307a430ecd18" exitCode=0 Dec 09 17:26:35 crc kubenswrapper[4954]: I1209 17:26:35.225732 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l6s5g-config-gn9nb" event={"ID":"559ebfa2-b84e-4b18-901d-314bdddd8fcc","Type":"ContainerDied","Data":"35b1cdfbacb3a5a691cd62bbe33d3dc479c80cc894a88016c2cf307a430ecd18"} Dec 09 17:26:35 crc kubenswrapper[4954]: I1209 17:26:35.237740 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4vtb6" Dec 09 17:26:35 crc kubenswrapper[4954]: I1209 17:26:35.238085 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4vtb6" event={"ID":"9fc44917-47c6-450b-b393-bad021024fcd","Type":"ContainerDied","Data":"9b8d18825561257eb7a7f77963ab63c8e59ecf69a4e7496fb1077a4c35dcf52f"} Dec 09 17:26:35 crc kubenswrapper[4954]: I1209 17:26:35.238190 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b8d18825561257eb7a7f77963ab63c8e59ecf69a4e7496fb1077a4c35dcf52f" Dec 09 17:26:35 crc kubenswrapper[4954]: I1209 17:26:35.240874 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:37 crc kubenswrapper[4954]: I1209 17:26:37.967261 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 17:26:37 crc kubenswrapper[4954]: I1209 17:26:37.968508 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="prometheus" containerID="cri-o://d691d5676e0e085b4b4d4e79b096c31307d777d05ac20dc56d385175663416e9" gracePeriod=600 Dec 09 17:26:37 crc kubenswrapper[4954]: I1209 17:26:37.968769 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="config-reloader" containerID="cri-o://8c9caea0a581dfbe0d2b6c0035f2f0087c3e7b00bb3b4c12dd285ab0ebe0e6b5" gracePeriod=600 Dec 09 17:26:37 crc kubenswrapper[4954]: I1209 17:26:37.968827 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="thanos-sidecar" containerID="cri-o://96f69de0abcceb205010d2c1050f29f7a03c6c5d48529e92e4467ff69964a002" gracePeriod=600 Dec 09 17:26:38 crc kubenswrapper[4954]: I1209 17:26:38.122199 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:26:38 crc kubenswrapper[4954]: E1209 17:26:38.122741 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:26:39 crc kubenswrapper[4954]: I1209 17:26:39.298696 4954 generic.go:334] "Generic (PLEG): container finished" podID="f0da1b73-3249-449b-86f9-12206ce93aab" containerID="96f69de0abcceb205010d2c1050f29f7a03c6c5d48529e92e4467ff69964a002" exitCode=0 Dec 09 17:26:39 crc kubenswrapper[4954]: I1209 17:26:39.299068 4954 generic.go:334] "Generic (PLEG): container finished" podID="f0da1b73-3249-449b-86f9-12206ce93aab" containerID="8c9caea0a581dfbe0d2b6c0035f2f0087c3e7b00bb3b4c12dd285ab0ebe0e6b5" exitCode=0 Dec 09 17:26:39 crc kubenswrapper[4954]: I1209 17:26:39.299085 4954 generic.go:334] "Generic (PLEG): container finished" podID="f0da1b73-3249-449b-86f9-12206ce93aab" containerID="d691d5676e0e085b4b4d4e79b096c31307d777d05ac20dc56d385175663416e9" exitCode=0 Dec 09 17:26:39 crc kubenswrapper[4954]: I1209 17:26:39.298919 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f0da1b73-3249-449b-86f9-12206ce93aab","Type":"ContainerDied","Data":"96f69de0abcceb205010d2c1050f29f7a03c6c5d48529e92e4467ff69964a002"} Dec 09 17:26:39 crc kubenswrapper[4954]: I1209 17:26:39.299134 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f0da1b73-3249-449b-86f9-12206ce93aab","Type":"ContainerDied","Data":"8c9caea0a581dfbe0d2b6c0035f2f0087c3e7b00bb3b4c12dd285ab0ebe0e6b5"} Dec 09 17:26:39 crc kubenswrapper[4954]: I1209 17:26:39.299154 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f0da1b73-3249-449b-86f9-12206ce93aab","Type":"ContainerDied","Data":"d691d5676e0e085b4b4d4e79b096c31307d777d05ac20dc56d385175663416e9"} Dec 09 17:26:39 crc kubenswrapper[4954]: I1209 17:26:39.799691 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.133:9090/-/ready\": dial tcp 10.217.0.133:9090: connect: connection refused" Dec 09 17:26:41 crc kubenswrapper[4954]: I1209 17:26:41.006060 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:41 crc kubenswrapper[4954]: I1209 17:26:41.014920 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69b248e7-6cb4-4805-a01f-537fe58aa42d-etc-swift\") pod \"swift-storage-0\" (UID: \"69b248e7-6cb4-4805-a01f-537fe58aa42d\") " pod="openstack/swift-storage-0" Dec 09 17:26:41 crc kubenswrapper[4954]: I1209 17:26:41.019570 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.316613 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.804465 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-mkljg"] Dec 09 17:26:42 crc kubenswrapper[4954]: E1209 17:26:42.805910 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc44917-47c6-450b-b393-bad021024fcd" containerName="swift-ring-rebalance" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.806011 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc44917-47c6-450b-b393-bad021024fcd" containerName="swift-ring-rebalance" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.806299 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fc44917-47c6-450b-b393-bad021024fcd" containerName="swift-ring-rebalance" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.807270 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-mkljg" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.821311 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-5d75-account-create-update-shqjl"] Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.823423 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5d75-account-create-update-shqjl" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.828617 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.840060 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-mkljg"] Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.892690 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-5d75-account-create-update-shqjl"] Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.921235 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-tnzh6"] Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.923132 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tnzh6" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.932065 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tnzh6"] Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.957156 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twx87\" (UniqueName: \"kubernetes.io/projected/bfcb15dd-bfcb-4519-917e-b41df5be28a9-kube-api-access-twx87\") pod \"cinder-db-create-tnzh6\" (UID: \"bfcb15dd-bfcb-4519-917e-b41df5be28a9\") " pod="openstack/cinder-db-create-tnzh6" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.957227 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfcb15dd-bfcb-4519-917e-b41df5be28a9-operator-scripts\") pod \"cinder-db-create-tnzh6\" (UID: \"bfcb15dd-bfcb-4519-917e-b41df5be28a9\") " pod="openstack/cinder-db-create-tnzh6" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.957276 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqzfj\" (UniqueName: \"kubernetes.io/projected/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-kube-api-access-hqzfj\") pod \"heat-db-create-mkljg\" (UID: \"1e02fe72-06d2-43b5-add1-f6577e3bcc6c\") " pod="openstack/heat-db-create-mkljg" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.957298 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c96j\" (UniqueName: \"kubernetes.io/projected/36f77ede-a607-4d03-b939-2e8908e27def-kube-api-access-8c96j\") pod \"heat-5d75-account-create-update-shqjl\" (UID: \"36f77ede-a607-4d03-b939-2e8908e27def\") " pod="openstack/heat-5d75-account-create-update-shqjl" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.957344 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-operator-scripts\") pod \"heat-db-create-mkljg\" (UID: \"1e02fe72-06d2-43b5-add1-f6577e3bcc6c\") " pod="openstack/heat-db-create-mkljg" Dec 09 17:26:42 crc kubenswrapper[4954]: I1209 17:26:42.957376 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36f77ede-a607-4d03-b939-2e8908e27def-operator-scripts\") pod \"heat-5d75-account-create-update-shqjl\" (UID: \"36f77ede-a607-4d03-b939-2e8908e27def\") " pod="openstack/heat-5d75-account-create-update-shqjl" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.059407 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfcb15dd-bfcb-4519-917e-b41df5be28a9-operator-scripts\") pod \"cinder-db-create-tnzh6\" (UID: \"bfcb15dd-bfcb-4519-917e-b41df5be28a9\") " pod="openstack/cinder-db-create-tnzh6" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.059507 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqzfj\" (UniqueName: \"kubernetes.io/projected/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-kube-api-access-hqzfj\") pod \"heat-db-create-mkljg\" (UID: \"1e02fe72-06d2-43b5-add1-f6577e3bcc6c\") " pod="openstack/heat-db-create-mkljg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.059542 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c96j\" (UniqueName: \"kubernetes.io/projected/36f77ede-a607-4d03-b939-2e8908e27def-kube-api-access-8c96j\") pod \"heat-5d75-account-create-update-shqjl\" (UID: \"36f77ede-a607-4d03-b939-2e8908e27def\") " pod="openstack/heat-5d75-account-create-update-shqjl" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.059638 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-operator-scripts\") pod \"heat-db-create-mkljg\" (UID: \"1e02fe72-06d2-43b5-add1-f6577e3bcc6c\") " pod="openstack/heat-db-create-mkljg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.059707 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36f77ede-a607-4d03-b939-2e8908e27def-operator-scripts\") pod \"heat-5d75-account-create-update-shqjl\" (UID: \"36f77ede-a607-4d03-b939-2e8908e27def\") " pod="openstack/heat-5d75-account-create-update-shqjl" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.059783 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twx87\" (UniqueName: \"kubernetes.io/projected/bfcb15dd-bfcb-4519-917e-b41df5be28a9-kube-api-access-twx87\") pod \"cinder-db-create-tnzh6\" (UID: \"bfcb15dd-bfcb-4519-917e-b41df5be28a9\") " pod="openstack/cinder-db-create-tnzh6" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.060368 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfcb15dd-bfcb-4519-917e-b41df5be28a9-operator-scripts\") pod \"cinder-db-create-tnzh6\" (UID: \"bfcb15dd-bfcb-4519-917e-b41df5be28a9\") " pod="openstack/cinder-db-create-tnzh6" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.061057 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-operator-scripts\") pod \"heat-db-create-mkljg\" (UID: \"1e02fe72-06d2-43b5-add1-f6577e3bcc6c\") " pod="openstack/heat-db-create-mkljg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.061173 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36f77ede-a607-4d03-b939-2e8908e27def-operator-scripts\") pod \"heat-5d75-account-create-update-shqjl\" (UID: \"36f77ede-a607-4d03-b939-2e8908e27def\") " pod="openstack/heat-5d75-account-create-update-shqjl" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.091456 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twx87\" (UniqueName: \"kubernetes.io/projected/bfcb15dd-bfcb-4519-917e-b41df5be28a9-kube-api-access-twx87\") pod \"cinder-db-create-tnzh6\" (UID: \"bfcb15dd-bfcb-4519-917e-b41df5be28a9\") " pod="openstack/cinder-db-create-tnzh6" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.098794 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c96j\" (UniqueName: \"kubernetes.io/projected/36f77ede-a607-4d03-b939-2e8908e27def-kube-api-access-8c96j\") pod \"heat-5d75-account-create-update-shqjl\" (UID: \"36f77ede-a607-4d03-b939-2e8908e27def\") " pod="openstack/heat-5d75-account-create-update-shqjl" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.101328 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqzfj\" (UniqueName: \"kubernetes.io/projected/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-kube-api-access-hqzfj\") pod \"heat-db-create-mkljg\" (UID: \"1e02fe72-06d2-43b5-add1-f6577e3bcc6c\") " pod="openstack/heat-db-create-mkljg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.113441 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-l48m5"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.124920 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-l48m5" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.145462 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-l48m5"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.153371 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-mkljg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.161871 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7vbk\" (UniqueName: \"kubernetes.io/projected/8c6703cc-4185-4dac-ab05-40f953f09e2e-kube-api-access-v7vbk\") pod \"barbican-db-create-l48m5\" (UID: \"8c6703cc-4185-4dac-ab05-40f953f09e2e\") " pod="openstack/barbican-db-create-l48m5" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.162006 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c6703cc-4185-4dac-ab05-40f953f09e2e-operator-scripts\") pod \"barbican-db-create-l48m5\" (UID: \"8c6703cc-4185-4dac-ab05-40f953f09e2e\") " pod="openstack/barbican-db-create-l48m5" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.173138 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b62d-account-create-update-2jx8d"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.178152 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5d75-account-create-update-shqjl" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.193781 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b62d-account-create-update-2jx8d" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.198152 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.241205 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b62d-account-create-update-2jx8d"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.260058 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tnzh6" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.273527 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c6703cc-4185-4dac-ab05-40f953f09e2e-operator-scripts\") pod \"barbican-db-create-l48m5\" (UID: \"8c6703cc-4185-4dac-ab05-40f953f09e2e\") " pod="openstack/barbican-db-create-l48m5" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.273701 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf8dz\" (UniqueName: \"kubernetes.io/projected/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-kube-api-access-jf8dz\") pod \"neutron-b62d-account-create-update-2jx8d\" (UID: \"d5b0ee6c-24bf-4b3f-875c-065c7f51862b\") " pod="openstack/neutron-b62d-account-create-update-2jx8d" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.274094 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7vbk\" (UniqueName: \"kubernetes.io/projected/8c6703cc-4185-4dac-ab05-40f953f09e2e-kube-api-access-v7vbk\") pod \"barbican-db-create-l48m5\" (UID: \"8c6703cc-4185-4dac-ab05-40f953f09e2e\") " pod="openstack/barbican-db-create-l48m5" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.274127 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-operator-scripts\") pod \"neutron-b62d-account-create-update-2jx8d\" (UID: \"d5b0ee6c-24bf-4b3f-875c-065c7f51862b\") " pod="openstack/neutron-b62d-account-create-update-2jx8d" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.279902 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c6703cc-4185-4dac-ab05-40f953f09e2e-operator-scripts\") pod \"barbican-db-create-l48m5\" (UID: \"8c6703cc-4185-4dac-ab05-40f953f09e2e\") " pod="openstack/barbican-db-create-l48m5" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.281831 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wq9lg"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.292881 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.301780 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.302287 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5xlkl" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.302931 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.304233 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wq9lg"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.304200 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.318973 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7vbk\" (UniqueName: \"kubernetes.io/projected/8c6703cc-4185-4dac-ab05-40f953f09e2e-kube-api-access-v7vbk\") pod \"barbican-db-create-l48m5\" (UID: \"8c6703cc-4185-4dac-ab05-40f953f09e2e\") " pod="openstack/barbican-db-create-l48m5" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.351138 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-51e0-account-create-update-krjpx"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.353816 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-51e0-account-create-update-krjpx" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.363039 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-51e0-account-create-update-krjpx"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.372050 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.372269 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-8m66z"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.373842 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8m66z" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.379023 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-operator-scripts\") pod \"neutron-b62d-account-create-update-2jx8d\" (UID: \"d5b0ee6c-24bf-4b3f-875c-065c7f51862b\") " pod="openstack/neutron-b62d-account-create-update-2jx8d" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.379136 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf8dz\" (UniqueName: \"kubernetes.io/projected/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-kube-api-access-jf8dz\") pod \"neutron-b62d-account-create-update-2jx8d\" (UID: \"d5b0ee6c-24bf-4b3f-875c-065c7f51862b\") " pod="openstack/neutron-b62d-account-create-update-2jx8d" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.380294 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-operator-scripts\") pod \"neutron-b62d-account-create-update-2jx8d\" (UID: \"d5b0ee6c-24bf-4b3f-875c-065c7f51862b\") " pod="openstack/neutron-b62d-account-create-update-2jx8d" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.392765 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-4af7-account-create-update-pmqz4"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.394308 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4af7-account-create-update-pmqz4" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.403478 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8m66z"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.411076 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.465492 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4af7-account-create-update-pmqz4"] Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.482362 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc2hq\" (UniqueName: \"kubernetes.io/projected/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-kube-api-access-fc2hq\") pod \"neutron-db-create-8m66z\" (UID: \"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74\") " pod="openstack/neutron-db-create-8m66z" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.482430 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-operator-scripts\") pod \"neutron-db-create-8m66z\" (UID: \"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74\") " pod="openstack/neutron-db-create-8m66z" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.482456 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-config-data\") pod \"keystone-db-sync-wq9lg\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.482517 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f32a6b81-b028-4e1d-b32b-2d8271d13875-operator-scripts\") pod \"barbican-51e0-account-create-update-krjpx\" (UID: \"f32a6b81-b028-4e1d-b32b-2d8271d13875\") " pod="openstack/barbican-51e0-account-create-update-krjpx" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.482548 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7llg\" (UniqueName: \"kubernetes.io/projected/f32a6b81-b028-4e1d-b32b-2d8271d13875-kube-api-access-d7llg\") pod \"barbican-51e0-account-create-update-krjpx\" (UID: \"f32a6b81-b028-4e1d-b32b-2d8271d13875\") " pod="openstack/barbican-51e0-account-create-update-krjpx" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.482607 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-combined-ca-bundle\") pod \"keystone-db-sync-wq9lg\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.482663 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4zz9\" (UniqueName: \"kubernetes.io/projected/b75d1898-5670-4f9e-88d9-69e4a1348c95-kube-api-access-z4zz9\") pod \"keystone-db-sync-wq9lg\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.496560 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf8dz\" (UniqueName: \"kubernetes.io/projected/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-kube-api-access-jf8dz\") pod \"neutron-b62d-account-create-update-2jx8d\" (UID: \"d5b0ee6c-24bf-4b3f-875c-065c7f51862b\") " pod="openstack/neutron-b62d-account-create-update-2jx8d" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.584345 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75288628-3722-4b4a-a76c-bd448cd5996c-operator-scripts\") pod \"cinder-4af7-account-create-update-pmqz4\" (UID: \"75288628-3722-4b4a-a76c-bd448cd5996c\") " pod="openstack/cinder-4af7-account-create-update-pmqz4" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.584480 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9jll\" (UniqueName: \"kubernetes.io/projected/75288628-3722-4b4a-a76c-bd448cd5996c-kube-api-access-d9jll\") pod \"cinder-4af7-account-create-update-pmqz4\" (UID: \"75288628-3722-4b4a-a76c-bd448cd5996c\") " pod="openstack/cinder-4af7-account-create-update-pmqz4" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.584520 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc2hq\" (UniqueName: \"kubernetes.io/projected/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-kube-api-access-fc2hq\") pod \"neutron-db-create-8m66z\" (UID: \"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74\") " pod="openstack/neutron-db-create-8m66z" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.584546 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-operator-scripts\") pod \"neutron-db-create-8m66z\" (UID: \"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74\") " pod="openstack/neutron-db-create-8m66z" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.584572 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-config-data\") pod \"keystone-db-sync-wq9lg\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.584627 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f32a6b81-b028-4e1d-b32b-2d8271d13875-operator-scripts\") pod \"barbican-51e0-account-create-update-krjpx\" (UID: \"f32a6b81-b028-4e1d-b32b-2d8271d13875\") " pod="openstack/barbican-51e0-account-create-update-krjpx" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.584656 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7llg\" (UniqueName: \"kubernetes.io/projected/f32a6b81-b028-4e1d-b32b-2d8271d13875-kube-api-access-d7llg\") pod \"barbican-51e0-account-create-update-krjpx\" (UID: \"f32a6b81-b028-4e1d-b32b-2d8271d13875\") " pod="openstack/barbican-51e0-account-create-update-krjpx" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.584701 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-combined-ca-bundle\") pod \"keystone-db-sync-wq9lg\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.584771 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4zz9\" (UniqueName: \"kubernetes.io/projected/b75d1898-5670-4f9e-88d9-69e4a1348c95-kube-api-access-z4zz9\") pod \"keystone-db-sync-wq9lg\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.586571 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-operator-scripts\") pod \"neutron-db-create-8m66z\" (UID: \"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74\") " pod="openstack/neutron-db-create-8m66z" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.588515 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f32a6b81-b028-4e1d-b32b-2d8271d13875-operator-scripts\") pod \"barbican-51e0-account-create-update-krjpx\" (UID: \"f32a6b81-b028-4e1d-b32b-2d8271d13875\") " pod="openstack/barbican-51e0-account-create-update-krjpx" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.592244 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-config-data\") pod \"keystone-db-sync-wq9lg\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.601873 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-combined-ca-bundle\") pod \"keystone-db-sync-wq9lg\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.605900 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7llg\" (UniqueName: \"kubernetes.io/projected/f32a6b81-b028-4e1d-b32b-2d8271d13875-kube-api-access-d7llg\") pod \"barbican-51e0-account-create-update-krjpx\" (UID: \"f32a6b81-b028-4e1d-b32b-2d8271d13875\") " pod="openstack/barbican-51e0-account-create-update-krjpx" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.608053 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-l48m5" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.609803 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4zz9\" (UniqueName: \"kubernetes.io/projected/b75d1898-5670-4f9e-88d9-69e4a1348c95-kube-api-access-z4zz9\") pod \"keystone-db-sync-wq9lg\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.613206 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc2hq\" (UniqueName: \"kubernetes.io/projected/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-kube-api-access-fc2hq\") pod \"neutron-db-create-8m66z\" (UID: \"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74\") " pod="openstack/neutron-db-create-8m66z" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.614155 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b62d-account-create-update-2jx8d" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.686349 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75288628-3722-4b4a-a76c-bd448cd5996c-operator-scripts\") pod \"cinder-4af7-account-create-update-pmqz4\" (UID: \"75288628-3722-4b4a-a76c-bd448cd5996c\") " pod="openstack/cinder-4af7-account-create-update-pmqz4" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.686496 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9jll\" (UniqueName: \"kubernetes.io/projected/75288628-3722-4b4a-a76c-bd448cd5996c-kube-api-access-d9jll\") pod \"cinder-4af7-account-create-update-pmqz4\" (UID: \"75288628-3722-4b4a-a76c-bd448cd5996c\") " pod="openstack/cinder-4af7-account-create-update-pmqz4" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.687350 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75288628-3722-4b4a-a76c-bd448cd5996c-operator-scripts\") pod \"cinder-4af7-account-create-update-pmqz4\" (UID: \"75288628-3722-4b4a-a76c-bd448cd5996c\") " pod="openstack/cinder-4af7-account-create-update-pmqz4" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.712982 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.719218 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9jll\" (UniqueName: \"kubernetes.io/projected/75288628-3722-4b4a-a76c-bd448cd5996c-kube-api-access-d9jll\") pod \"cinder-4af7-account-create-update-pmqz4\" (UID: \"75288628-3722-4b4a-a76c-bd448cd5996c\") " pod="openstack/cinder-4af7-account-create-update-pmqz4" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.745287 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-51e0-account-create-update-krjpx" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.758276 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8m66z" Dec 09 17:26:43 crc kubenswrapper[4954]: I1209 17:26:43.772862 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4af7-account-create-update-pmqz4" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.446608 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l6s5g-config-gn9nb" event={"ID":"559ebfa2-b84e-4b18-901d-314bdddd8fcc","Type":"ContainerDied","Data":"ea39cf07b2eb95c699b2f29b4dd245d16adcebea30009013e7cd484c04892421"} Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.446975 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea39cf07b2eb95c699b2f29b4dd245d16adcebea30009013e7cd484c04892421" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.478967 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.611453 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run\") pod \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.611893 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-additional-scripts\") pod \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.611952 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run-ovn\") pod \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.611625 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run" (OuterVolumeSpecName: "var-run") pod "559ebfa2-b84e-4b18-901d-314bdddd8fcc" (UID: "559ebfa2-b84e-4b18-901d-314bdddd8fcc"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.612034 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "559ebfa2-b84e-4b18-901d-314bdddd8fcc" (UID: "559ebfa2-b84e-4b18-901d-314bdddd8fcc"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.612426 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "559ebfa2-b84e-4b18-901d-314bdddd8fcc" (UID: "559ebfa2-b84e-4b18-901d-314bdddd8fcc"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.612557 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-log-ovn\") pod \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.612675 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz8pm\" (UniqueName: \"kubernetes.io/projected/559ebfa2-b84e-4b18-901d-314bdddd8fcc-kube-api-access-rz8pm\") pod \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.612719 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-scripts\") pod \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\" (UID: \"559ebfa2-b84e-4b18-901d-314bdddd8fcc\") " Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.613214 4954 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.613227 4954 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.613238 4954 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.614447 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "559ebfa2-b84e-4b18-901d-314bdddd8fcc" (UID: "559ebfa2-b84e-4b18-901d-314bdddd8fcc"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.619381 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/559ebfa2-b84e-4b18-901d-314bdddd8fcc-kube-api-access-rz8pm" (OuterVolumeSpecName: "kube-api-access-rz8pm") pod "559ebfa2-b84e-4b18-901d-314bdddd8fcc" (UID: "559ebfa2-b84e-4b18-901d-314bdddd8fcc"). InnerVolumeSpecName "kube-api-access-rz8pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.619720 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-scripts" (OuterVolumeSpecName: "scripts") pod "559ebfa2-b84e-4b18-901d-314bdddd8fcc" (UID: "559ebfa2-b84e-4b18-901d-314bdddd8fcc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.725359 4954 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/559ebfa2-b84e-4b18-901d-314bdddd8fcc-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.725411 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz8pm\" (UniqueName: \"kubernetes.io/projected/559ebfa2-b84e-4b18-901d-314bdddd8fcc-kube-api-access-rz8pm\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:44 crc kubenswrapper[4954]: I1209 17:26:44.725427 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/559ebfa2-b84e-4b18-901d-314bdddd8fcc-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.016631 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.055020 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f0da1b73-3249-449b-86f9-12206ce93aab-prometheus-metric-storage-rulefiles-0\") pod \"f0da1b73-3249-449b-86f9-12206ce93aab\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.055080 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-tls-assets\") pod \"f0da1b73-3249-449b-86f9-12206ce93aab\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.055125 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f0da1b73-3249-449b-86f9-12206ce93aab-config-out\") pod \"f0da1b73-3249-449b-86f9-12206ce93aab\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.055173 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"f0da1b73-3249-449b-86f9-12206ce93aab\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.055260 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-config\") pod \"f0da1b73-3249-449b-86f9-12206ce93aab\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.055316 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-web-config\") pod \"f0da1b73-3249-449b-86f9-12206ce93aab\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.055458 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w4n7\" (UniqueName: \"kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-kube-api-access-8w4n7\") pod \"f0da1b73-3249-449b-86f9-12206ce93aab\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.055508 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-thanos-prometheus-http-client-file\") pod \"f0da1b73-3249-449b-86f9-12206ce93aab\" (UID: \"f0da1b73-3249-449b-86f9-12206ce93aab\") " Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.059754 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0da1b73-3249-449b-86f9-12206ce93aab-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "f0da1b73-3249-449b-86f9-12206ce93aab" (UID: "f0da1b73-3249-449b-86f9-12206ce93aab"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.102684 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "f0da1b73-3249-449b-86f9-12206ce93aab" (UID: "f0da1b73-3249-449b-86f9-12206ce93aab"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.138757 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0da1b73-3249-449b-86f9-12206ce93aab-config-out" (OuterVolumeSpecName: "config-out") pod "f0da1b73-3249-449b-86f9-12206ce93aab" (UID: "f0da1b73-3249-449b-86f9-12206ce93aab"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.138878 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-config" (OuterVolumeSpecName: "config") pod "f0da1b73-3249-449b-86f9-12206ce93aab" (UID: "f0da1b73-3249-449b-86f9-12206ce93aab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.139191 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "f0da1b73-3249-449b-86f9-12206ce93aab" (UID: "f0da1b73-3249-449b-86f9-12206ce93aab"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.143889 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-kube-api-access-8w4n7" (OuterVolumeSpecName: "kube-api-access-8w4n7") pod "f0da1b73-3249-449b-86f9-12206ce93aab" (UID: "f0da1b73-3249-449b-86f9-12206ce93aab"). InnerVolumeSpecName "kube-api-access-8w4n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.157977 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w4n7\" (UniqueName: \"kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-kube-api-access-8w4n7\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.158010 4954 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.158027 4954 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f0da1b73-3249-449b-86f9-12206ce93aab-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.158036 4954 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f0da1b73-3249-449b-86f9-12206ce93aab-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.158045 4954 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f0da1b73-3249-449b-86f9-12206ce93aab-config-out\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.158056 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.185939 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-web-config" (OuterVolumeSpecName: "web-config") pod "f0da1b73-3249-449b-86f9-12206ce93aab" (UID: "f0da1b73-3249-449b-86f9-12206ce93aab"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.195492 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "f0da1b73-3249-449b-86f9-12206ce93aab" (UID: "f0da1b73-3249-449b-86f9-12206ce93aab"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.263026 4954 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.263059 4954 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f0da1b73-3249-449b-86f9-12206ce93aab-web-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.290761 4954 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.365059 4954 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.469692 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l6s5g-config-gn9nb" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.474865 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f0da1b73-3249-449b-86f9-12206ce93aab","Type":"ContainerDied","Data":"53ec0de08d9d6b12288f1fce35364fa6c66400e75e8c152bf38b44f97a107ca5"} Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.474947 4954 scope.go:117] "RemoveContainer" containerID="96f69de0abcceb205010d2c1050f29f7a03c6c5d48529e92e4467ff69964a002" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.474950 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.542692 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.629567 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.660844 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 17:26:45 crc kubenswrapper[4954]: E1209 17:26:45.661542 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="config-reloader" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.661565 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="config-reloader" Dec 09 17:26:45 crc kubenswrapper[4954]: E1209 17:26:45.661671 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="thanos-sidecar" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.661701 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="thanos-sidecar" Dec 09 17:26:45 crc kubenswrapper[4954]: E1209 17:26:45.661712 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="prometheus" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.661721 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="prometheus" Dec 09 17:26:45 crc kubenswrapper[4954]: E1209 17:26:45.661735 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559ebfa2-b84e-4b18-901d-314bdddd8fcc" containerName="ovn-config" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.661743 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="559ebfa2-b84e-4b18-901d-314bdddd8fcc" containerName="ovn-config" Dec 09 17:26:45 crc kubenswrapper[4954]: E1209 17:26:45.661795 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="init-config-reloader" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.661803 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="init-config-reloader" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.662316 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="559ebfa2-b84e-4b18-901d-314bdddd8fcc" containerName="ovn-config" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.662371 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="prometheus" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.662484 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="thanos-sidecar" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.662497 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="config-reloader" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.667010 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.670809 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-xwpb8" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.672093 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.672283 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.672482 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.672742 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.674465 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.689829 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.692540 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.731685 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-l6s5g-config-gn9nb"] Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.743726 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-l6s5g-config-gn9nb"] Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.755763 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-mkljg"] Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792106 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792191 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792219 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj79j\" (UniqueName: \"kubernetes.io/projected/e626f486-ed0d-4388-9040-1c23a41aba18-kube-api-access-mj79j\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792244 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-config\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792269 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792336 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792358 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792385 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792407 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e626f486-ed0d-4388-9040-1c23a41aba18-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792453 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e626f486-ed0d-4388-9040-1c23a41aba18-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.792473 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e626f486-ed0d-4388-9040-1c23a41aba18-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.868973 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-51e0-account-create-update-krjpx"] Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.895849 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.896197 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj79j\" (UniqueName: \"kubernetes.io/projected/e626f486-ed0d-4388-9040-1c23a41aba18-kube-api-access-mj79j\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.896291 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-config\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.896454 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.896862 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.897095 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.897215 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.897315 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e626f486-ed0d-4388-9040-1c23a41aba18-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.897511 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e626f486-ed0d-4388-9040-1c23a41aba18-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.902776 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e626f486-ed0d-4388-9040-1c23a41aba18-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.897789 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.903807 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e626f486-ed0d-4388-9040-1c23a41aba18-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.906743 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.907144 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e626f486-ed0d-4388-9040-1c23a41aba18-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.907365 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.908271 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-config\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.910681 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.910969 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e626f486-ed0d-4388-9040-1c23a41aba18-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.920698 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.920729 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.922550 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e626f486-ed0d-4388-9040-1c23a41aba18-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.938510 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj79j\" (UniqueName: \"kubernetes.io/projected/e626f486-ed0d-4388-9040-1c23a41aba18-kube-api-access-mj79j\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:45 crc kubenswrapper[4954]: I1209 17:26:45.960810 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"e626f486-ed0d-4388-9040-1c23a41aba18\") " pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.002171 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.061791 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b62d-account-create-update-2jx8d"] Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.071200 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tnzh6"] Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.139826 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="559ebfa2-b84e-4b18-901d-314bdddd8fcc" path="/var/lib/kubelet/pods/559ebfa2-b84e-4b18-901d-314bdddd8fcc/volumes" Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.148003 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" path="/var/lib/kubelet/pods/f0da1b73-3249-449b-86f9-12206ce93aab/volumes" Dec 09 17:26:46 crc kubenswrapper[4954]: W1209 17:26:46.217177 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e02fe72_06d2_43b5_add1_f6577e3bcc6c.slice/crio-3d9bf6f753989e154acd35e99393db6a7c853925d8697aa714ea2130218f5a11 WatchSource:0}: Error finding container 3d9bf6f753989e154acd35e99393db6a7c853925d8697aa714ea2130218f5a11: Status 404 returned error can't find the container with id 3d9bf6f753989e154acd35e99393db6a7c853925d8697aa714ea2130218f5a11 Dec 09 17:26:46 crc kubenswrapper[4954]: W1209 17:26:46.218855 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf32a6b81_b028_4e1d_b32b_2d8271d13875.slice/crio-acb5de494aad3a607c058fedb05a70dbc902450a60d3824ef8e09e69d1a764fc WatchSource:0}: Error finding container acb5de494aad3a607c058fedb05a70dbc902450a60d3824ef8e09e69d1a764fc: Status 404 returned error can't find the container with id acb5de494aad3a607c058fedb05a70dbc902450a60d3824ef8e09e69d1a764fc Dec 09 17:26:46 crc kubenswrapper[4954]: W1209 17:26:46.250198 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5b0ee6c_24bf_4b3f_875c_065c7f51862b.slice/crio-012ac8c449e38b12719392aa2df33a21f6a78444ad56e2b1ba4656d6e205cadb WatchSource:0}: Error finding container 012ac8c449e38b12719392aa2df33a21f6a78444ad56e2b1ba4656d6e205cadb: Status 404 returned error can't find the container with id 012ac8c449e38b12719392aa2df33a21f6a78444ad56e2b1ba4656d6e205cadb Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.305119 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-l48m5"] Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.343796 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4af7-account-create-update-pmqz4"] Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.377835 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8m66z"] Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.402379 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wq9lg"] Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.411146 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-5d75-account-create-update-shqjl"] Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.439051 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.487645 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tnzh6" event={"ID":"bfcb15dd-bfcb-4519-917e-b41df5be28a9","Type":"ContainerStarted","Data":"45b299bbb21d9445dd93e38cc30ec10f1fff11d6a2ec6deb5f45ac5a5edf6d2d"} Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.490120 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-mkljg" event={"ID":"1e02fe72-06d2-43b5-add1-f6577e3bcc6c","Type":"ContainerStarted","Data":"3d9bf6f753989e154acd35e99393db6a7c853925d8697aa714ea2130218f5a11"} Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.492285 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b62d-account-create-update-2jx8d" event={"ID":"d5b0ee6c-24bf-4b3f-875c-065c7f51862b","Type":"ContainerStarted","Data":"012ac8c449e38b12719392aa2df33a21f6a78444ad56e2b1ba4656d6e205cadb"} Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.494409 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-51e0-account-create-update-krjpx" event={"ID":"f32a6b81-b028-4e1d-b32b-2d8271d13875","Type":"ContainerStarted","Data":"acb5de494aad3a607c058fedb05a70dbc902450a60d3824ef8e09e69d1a764fc"} Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.642942 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.643149 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 09 17:26:46 crc kubenswrapper[4954]: W1209 17:26:46.651508 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75288628_3722_4b4a_a76c_bd448cd5996c.slice/crio-55237dc6a6385a83360286012a80f7a53addae4c5074a8817a21e071fb42fc2d WatchSource:0}: Error finding container 55237dc6a6385a83360286012a80f7a53addae4c5074a8817a21e071fb42fc2d: Status 404 returned error can't find the container with id 55237dc6a6385a83360286012a80f7a53addae4c5074a8817a21e071fb42fc2d Dec 09 17:26:46 crc kubenswrapper[4954]: W1209 17:26:46.652820 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ec8214b_cbe8_4ddf_ac1e_b3e8d5d1ac74.slice/crio-69e5ab2937189b3c874b8def2aa3a30b14cd315e49c1f9eac1232017104c9bec WatchSource:0}: Error finding container 69e5ab2937189b3c874b8def2aa3a30b14cd315e49c1f9eac1232017104c9bec: Status 404 returned error can't find the container with id 69e5ab2937189b3c874b8def2aa3a30b14cd315e49c1f9eac1232017104c9bec Dec 09 17:26:46 crc kubenswrapper[4954]: W1209 17:26:46.657636 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb75d1898_5670_4f9e_88d9_69e4a1348c95.slice/crio-632ef31dec5aea84c1a4c71791de6c946aec0269c84229f41c3cb2b16e862beb WatchSource:0}: Error finding container 632ef31dec5aea84c1a4c71791de6c946aec0269c84229f41c3cb2b16e862beb: Status 404 returned error can't find the container with id 632ef31dec5aea84c1a4c71791de6c946aec0269c84229f41c3cb2b16e862beb Dec 09 17:26:46 crc kubenswrapper[4954]: W1209 17:26:46.657888 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36f77ede_a607_4d03_b939_2e8908e27def.slice/crio-3db06ca0ef6343debb1debb013b16afe6c2c0f84cf33110040750a2835a62155 WatchSource:0}: Error finding container 3db06ca0ef6343debb1debb013b16afe6c2c0f84cf33110040750a2835a62155: Status 404 returned error can't find the container with id 3db06ca0ef6343debb1debb013b16afe6c2c0f84cf33110040750a2835a62155 Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.685391 4954 scope.go:117] "RemoveContainer" containerID="8c9caea0a581dfbe0d2b6c0035f2f0087c3e7b00bb3b4c12dd285ab0ebe0e6b5" Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.702146 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.748626 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.766005 4954 scope.go:117] "RemoveContainer" containerID="d691d5676e0e085b4b4d4e79b096c31307d777d05ac20dc56d385175663416e9" Dec 09 17:26:46 crc kubenswrapper[4954]: I1209 17:26:46.945024 4954 scope.go:117] "RemoveContainer" containerID="bfa38e1cf3605a9d66bcda5a2cb2ed06a8bfbea9eb7f3287ed2e14d8d3509a38" Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.287398 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.524084 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4af7-account-create-update-pmqz4" event={"ID":"75288628-3722-4b4a-a76c-bd448cd5996c","Type":"ContainerStarted","Data":"8a70d9a3e5ab03784ccf64ebf5ffcb86fa80998a545ca58f6941e2039ba6b627"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.524144 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4af7-account-create-update-pmqz4" event={"ID":"75288628-3722-4b4a-a76c-bd448cd5996c","Type":"ContainerStarted","Data":"55237dc6a6385a83360286012a80f7a53addae4c5074a8817a21e071fb42fc2d"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.532320 4954 generic.go:334] "Generic (PLEG): container finished" podID="bfcb15dd-bfcb-4519-917e-b41df5be28a9" containerID="77d1027a9b3778aa9e7f775fd239898d1f2e67f789ec6871e65de66c17f05eb0" exitCode=0 Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.532418 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tnzh6" event={"ID":"bfcb15dd-bfcb-4519-917e-b41df5be28a9","Type":"ContainerDied","Data":"77d1027a9b3778aa9e7f775fd239898d1f2e67f789ec6871e65de66c17f05eb0"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.538221 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-l48m5" event={"ID":"8c6703cc-4185-4dac-ab05-40f953f09e2e","Type":"ContainerStarted","Data":"46bd1e0d8bdb3d52b824343c9b64b4c76bd9290fcee7eb08a2b31393f1461bac"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.538278 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-l48m5" event={"ID":"8c6703cc-4185-4dac-ab05-40f953f09e2e","Type":"ContainerStarted","Data":"df3a3690f3c6be4e63192052e3cbcb4926065ad50428967cce32acc4710c3bcd"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.547049 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-mkljg" event={"ID":"1e02fe72-06d2-43b5-add1-f6577e3bcc6c","Type":"ContainerStarted","Data":"746184e81b7f4df22f167becd076855f55e80b82a88c67b894c381c978b3d0b5"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.548868 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-4af7-account-create-update-pmqz4" podStartSLOduration=4.548838634 podStartE2EDuration="4.548838634s" podCreationTimestamp="2025-12-09 17:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:47.543688663 +0000 UTC m=+1803.931862503" watchObservedRunningTime="2025-12-09 17:26:47.548838634 +0000 UTC m=+1803.937012454" Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.552182 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8m66z" event={"ID":"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74","Type":"ContainerStarted","Data":"4a158212a46076db0596674f8ee27ab9ddc5463498b0dc80de8a368ee4cb82d2"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.552249 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8m66z" event={"ID":"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74","Type":"ContainerStarted","Data":"69e5ab2937189b3c874b8def2aa3a30b14cd315e49c1f9eac1232017104c9bec"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.562006 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e626f486-ed0d-4388-9040-1c23a41aba18","Type":"ContainerStarted","Data":"0fb0e4f345cbf91739ebac3ceb112043fc46e15d8607af371aa9779998448466"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.564693 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-5d75-account-create-update-shqjl" event={"ID":"36f77ede-a607-4d03-b939-2e8908e27def","Type":"ContainerStarted","Data":"fd9533e90f600c7b5adf3df73dca386043e8140bad943067cb92af7829fa878f"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.564749 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-5d75-account-create-update-shqjl" event={"ID":"36f77ede-a607-4d03-b939-2e8908e27def","Type":"ContainerStarted","Data":"3db06ca0ef6343debb1debb013b16afe6c2c0f84cf33110040750a2835a62155"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.567475 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wq9lg" event={"ID":"b75d1898-5670-4f9e-88d9-69e4a1348c95","Type":"ContainerStarted","Data":"632ef31dec5aea84c1a4c71791de6c946aec0269c84229f41c3cb2b16e862beb"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.594247 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"6ce7e263cb94be6830a99849500d6fa0c567116b33fdf288211143a2126f2cd9"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.597657 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-l48m5" podStartSLOduration=4.597631591 podStartE2EDuration="4.597631591s" podCreationTimestamp="2025-12-09 17:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:47.590488098 +0000 UTC m=+1803.978661918" watchObservedRunningTime="2025-12-09 17:26:47.597631591 +0000 UTC m=+1803.985805411" Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.598255 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g4kjd" event={"ID":"8640fe28-2072-49f7-8c8d-35a2ea5432e0","Type":"ContainerStarted","Data":"eab74493a154b3368fc126aa75e9d1753a9220c6bf6879cabc7ff20617476208"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.601442 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b62d-account-create-update-2jx8d" event={"ID":"d5b0ee6c-24bf-4b3f-875c-065c7f51862b","Type":"ContainerStarted","Data":"b23fc88a6f35a9ec77252dc7911681f4bb0b40fb65b51e924b5b61dc0ffe0f51"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.604993 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-51e0-account-create-update-krjpx" event={"ID":"f32a6b81-b028-4e1d-b32b-2d8271d13875","Type":"ContainerStarted","Data":"9a0a0d8331b1e11d765c8c42cef5cbf36ba8a024bf35353a76a4ede6422ab865"} Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.632913 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-5d75-account-create-update-shqjl" podStartSLOduration=5.632889046 podStartE2EDuration="5.632889046s" podCreationTimestamp="2025-12-09 17:26:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:47.60971502 +0000 UTC m=+1803.997888840" watchObservedRunningTime="2025-12-09 17:26:47.632889046 +0000 UTC m=+1804.021062866" Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.637761 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-8m66z" podStartSLOduration=4.637750198 podStartE2EDuration="4.637750198s" podCreationTimestamp="2025-12-09 17:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:47.623411869 +0000 UTC m=+1804.011585709" watchObservedRunningTime="2025-12-09 17:26:47.637750198 +0000 UTC m=+1804.025924018" Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.652784 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-mkljg" podStartSLOduration=5.652764448 podStartE2EDuration="5.652764448s" podCreationTimestamp="2025-12-09 17:26:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:47.641177935 +0000 UTC m=+1804.029351755" watchObservedRunningTime="2025-12-09 17:26:47.652764448 +0000 UTC m=+1804.040938268" Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.676511 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-51e0-account-create-update-krjpx" podStartSLOduration=4.67648695 podStartE2EDuration="4.67648695s" podCreationTimestamp="2025-12-09 17:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:47.662026167 +0000 UTC m=+1804.050200007" watchObservedRunningTime="2025-12-09 17:26:47.67648695 +0000 UTC m=+1804.064660770" Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.688786 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-b62d-account-create-update-2jx8d" podStartSLOduration=4.688761175 podStartE2EDuration="4.688761175s" podCreationTimestamp="2025-12-09 17:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:26:47.680265509 +0000 UTC m=+1804.068439339" watchObservedRunningTime="2025-12-09 17:26:47.688761175 +0000 UTC m=+1804.076934995" Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.699781 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-g4kjd" podStartSLOduration=4.921316738 podStartE2EDuration="22.699756829s" podCreationTimestamp="2025-12-09 17:26:25 +0000 UTC" firstStartedPulling="2025-12-09 17:26:26.776409885 +0000 UTC m=+1783.164583705" lastFinishedPulling="2025-12-09 17:26:44.554849976 +0000 UTC m=+1800.943023796" observedRunningTime="2025-12-09 17:26:47.698908062 +0000 UTC m=+1804.087081882" watchObservedRunningTime="2025-12-09 17:26:47.699756829 +0000 UTC m=+1804.087930649" Dec 09 17:26:47 crc kubenswrapper[4954]: I1209 17:26:47.800570 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="f0da1b73-3249-449b-86f9-12206ce93aab" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.133:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.619454 4954 generic.go:334] "Generic (PLEG): container finished" podID="d5b0ee6c-24bf-4b3f-875c-065c7f51862b" containerID="b23fc88a6f35a9ec77252dc7911681f4bb0b40fb65b51e924b5b61dc0ffe0f51" exitCode=0 Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.619583 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b62d-account-create-update-2jx8d" event={"ID":"d5b0ee6c-24bf-4b3f-875c-065c7f51862b","Type":"ContainerDied","Data":"b23fc88a6f35a9ec77252dc7911681f4bb0b40fb65b51e924b5b61dc0ffe0f51"} Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.623195 4954 generic.go:334] "Generic (PLEG): container finished" podID="8c6703cc-4185-4dac-ab05-40f953f09e2e" containerID="46bd1e0d8bdb3d52b824343c9b64b4c76bd9290fcee7eb08a2b31393f1461bac" exitCode=0 Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.623289 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-l48m5" event={"ID":"8c6703cc-4185-4dac-ab05-40f953f09e2e","Type":"ContainerDied","Data":"46bd1e0d8bdb3d52b824343c9b64b4c76bd9290fcee7eb08a2b31393f1461bac"} Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.628025 4954 generic.go:334] "Generic (PLEG): container finished" podID="1e02fe72-06d2-43b5-add1-f6577e3bcc6c" containerID="746184e81b7f4df22f167becd076855f55e80b82a88c67b894c381c978b3d0b5" exitCode=0 Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.628122 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-mkljg" event={"ID":"1e02fe72-06d2-43b5-add1-f6577e3bcc6c","Type":"ContainerDied","Data":"746184e81b7f4df22f167becd076855f55e80b82a88c67b894c381c978b3d0b5"} Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.637286 4954 generic.go:334] "Generic (PLEG): container finished" podID="36f77ede-a607-4d03-b939-2e8908e27def" containerID="fd9533e90f600c7b5adf3df73dca386043e8140bad943067cb92af7829fa878f" exitCode=0 Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.637382 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-5d75-account-create-update-shqjl" event={"ID":"36f77ede-a607-4d03-b939-2e8908e27def","Type":"ContainerDied","Data":"fd9533e90f600c7b5adf3df73dca386043e8140bad943067cb92af7829fa878f"} Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.656189 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"ce89f21f-56e2-4107-ae7a-cb38603a2e01","Type":"ContainerStarted","Data":"7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4"} Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.668087 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"28f0872bd163a08a6575d864f5cf7f58280e4fc77238bb392bb61ff62428bbe0"} Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.670664 4954 generic.go:334] "Generic (PLEG): container finished" podID="6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74" containerID="4a158212a46076db0596674f8ee27ab9ddc5463498b0dc80de8a368ee4cb82d2" exitCode=0 Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.670779 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8m66z" event={"ID":"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74","Type":"ContainerDied","Data":"4a158212a46076db0596674f8ee27ab9ddc5463498b0dc80de8a368ee4cb82d2"} Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.680817 4954 generic.go:334] "Generic (PLEG): container finished" podID="75288628-3722-4b4a-a76c-bd448cd5996c" containerID="8a70d9a3e5ab03784ccf64ebf5ffcb86fa80998a545ca58f6941e2039ba6b627" exitCode=0 Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.681043 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4af7-account-create-update-pmqz4" event={"ID":"75288628-3722-4b4a-a76c-bd448cd5996c","Type":"ContainerDied","Data":"8a70d9a3e5ab03784ccf64ebf5ffcb86fa80998a545ca58f6941e2039ba6b627"} Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.688587 4954 generic.go:334] "Generic (PLEG): container finished" podID="f32a6b81-b028-4e1d-b32b-2d8271d13875" containerID="9a0a0d8331b1e11d765c8c42cef5cbf36ba8a024bf35353a76a4ede6422ab865" exitCode=0 Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.689603 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-51e0-account-create-update-krjpx" event={"ID":"f32a6b81-b028-4e1d-b32b-2d8271d13875","Type":"ContainerDied","Data":"9a0a0d8331b1e11d765c8c42cef5cbf36ba8a024bf35353a76a4ede6422ab865"} Dec 09 17:26:48 crc kubenswrapper[4954]: I1209 17:26:48.748873 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=3.7163240379999998 podStartE2EDuration="15.748848751s" podCreationTimestamp="2025-12-09 17:26:33 +0000 UTC" firstStartedPulling="2025-12-09 17:26:34.814985155 +0000 UTC m=+1791.203158975" lastFinishedPulling="2025-12-09 17:26:46.847509868 +0000 UTC m=+1803.235683688" observedRunningTime="2025-12-09 17:26:48.736213856 +0000 UTC m=+1805.124387676" watchObservedRunningTime="2025-12-09 17:26:48.748848751 +0000 UTC m=+1805.137022571" Dec 09 17:26:49 crc kubenswrapper[4954]: I1209 17:26:49.120491 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:26:49 crc kubenswrapper[4954]: E1209 17:26:49.121237 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:26:49 crc kubenswrapper[4954]: I1209 17:26:49.704822 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"d4194945da87e713be7b300c5d9e5738362c2290a6c72423d765d36e3bb6d931"} Dec 09 17:26:51 crc kubenswrapper[4954]: I1209 17:26:51.727356 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e626f486-ed0d-4388-9040-1c23a41aba18","Type":"ContainerStarted","Data":"5aa994a812fd1b2f37e9c788c7399d668ca1c6caaa68ebc22a42545ef121f6d0"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.115124 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tnzh6" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.129756 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5d75-account-create-update-shqjl" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.149684 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4af7-account-create-update-pmqz4" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.163212 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-l48m5" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.172971 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b62d-account-create-update-2jx8d" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.188903 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8m66z" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.200410 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-mkljg" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.202983 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfcb15dd-bfcb-4519-917e-b41df5be28a9-operator-scripts\") pod \"bfcb15dd-bfcb-4519-917e-b41df5be28a9\" (UID: \"bfcb15dd-bfcb-4519-917e-b41df5be28a9\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.207966 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twx87\" (UniqueName: \"kubernetes.io/projected/bfcb15dd-bfcb-4519-917e-b41df5be28a9-kube-api-access-twx87\") pod \"bfcb15dd-bfcb-4519-917e-b41df5be28a9\" (UID: \"bfcb15dd-bfcb-4519-917e-b41df5be28a9\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.209751 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfcb15dd-bfcb-4519-917e-b41df5be28a9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bfcb15dd-bfcb-4519-917e-b41df5be28a9" (UID: "bfcb15dd-bfcb-4519-917e-b41df5be28a9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.217952 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfcb15dd-bfcb-4519-917e-b41df5be28a9-kube-api-access-twx87" (OuterVolumeSpecName: "kube-api-access-twx87") pod "bfcb15dd-bfcb-4519-917e-b41df5be28a9" (UID: "bfcb15dd-bfcb-4519-917e-b41df5be28a9"). InnerVolumeSpecName "kube-api-access-twx87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.236470 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-51e0-account-create-update-krjpx" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.309620 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36f77ede-a607-4d03-b939-2e8908e27def-operator-scripts\") pod \"36f77ede-a607-4d03-b939-2e8908e27def\" (UID: \"36f77ede-a607-4d03-b939-2e8908e27def\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.309706 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf8dz\" (UniqueName: \"kubernetes.io/projected/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-kube-api-access-jf8dz\") pod \"d5b0ee6c-24bf-4b3f-875c-065c7f51862b\" (UID: \"d5b0ee6c-24bf-4b3f-875c-065c7f51862b\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.309825 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9jll\" (UniqueName: \"kubernetes.io/projected/75288628-3722-4b4a-a76c-bd448cd5996c-kube-api-access-d9jll\") pod \"75288628-3722-4b4a-a76c-bd448cd5996c\" (UID: \"75288628-3722-4b4a-a76c-bd448cd5996c\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.309871 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c96j\" (UniqueName: \"kubernetes.io/projected/36f77ede-a607-4d03-b939-2e8908e27def-kube-api-access-8c96j\") pod \"36f77ede-a607-4d03-b939-2e8908e27def\" (UID: \"36f77ede-a607-4d03-b939-2e8908e27def\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.309948 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7vbk\" (UniqueName: \"kubernetes.io/projected/8c6703cc-4185-4dac-ab05-40f953f09e2e-kube-api-access-v7vbk\") pod \"8c6703cc-4185-4dac-ab05-40f953f09e2e\" (UID: \"8c6703cc-4185-4dac-ab05-40f953f09e2e\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.309984 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75288628-3722-4b4a-a76c-bd448cd5996c-operator-scripts\") pod \"75288628-3722-4b4a-a76c-bd448cd5996c\" (UID: \"75288628-3722-4b4a-a76c-bd448cd5996c\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.310023 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c6703cc-4185-4dac-ab05-40f953f09e2e-operator-scripts\") pod \"8c6703cc-4185-4dac-ab05-40f953f09e2e\" (UID: \"8c6703cc-4185-4dac-ab05-40f953f09e2e\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.310050 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqzfj\" (UniqueName: \"kubernetes.io/projected/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-kube-api-access-hqzfj\") pod \"1e02fe72-06d2-43b5-add1-f6577e3bcc6c\" (UID: \"1e02fe72-06d2-43b5-add1-f6577e3bcc6c\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.310114 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-operator-scripts\") pod \"d5b0ee6c-24bf-4b3f-875c-065c7f51862b\" (UID: \"d5b0ee6c-24bf-4b3f-875c-065c7f51862b\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.310156 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-operator-scripts\") pod \"1e02fe72-06d2-43b5-add1-f6577e3bcc6c\" (UID: \"1e02fe72-06d2-43b5-add1-f6577e3bcc6c\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.310246 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc2hq\" (UniqueName: \"kubernetes.io/projected/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-kube-api-access-fc2hq\") pod \"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74\" (UID: \"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.310272 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-operator-scripts\") pod \"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74\" (UID: \"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.310840 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bfcb15dd-bfcb-4519-917e-b41df5be28a9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.310869 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twx87\" (UniqueName: \"kubernetes.io/projected/bfcb15dd-bfcb-4519-917e-b41df5be28a9-kube-api-access-twx87\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.310851 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36f77ede-a607-4d03-b939-2e8908e27def-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "36f77ede-a607-4d03-b939-2e8908e27def" (UID: "36f77ede-a607-4d03-b939-2e8908e27def"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.310955 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d5b0ee6c-24bf-4b3f-875c-065c7f51862b" (UID: "d5b0ee6c-24bf-4b3f-875c-065c7f51862b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.311112 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75288628-3722-4b4a-a76c-bd448cd5996c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "75288628-3722-4b4a-a76c-bd448cd5996c" (UID: "75288628-3722-4b4a-a76c-bd448cd5996c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.311420 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c6703cc-4185-4dac-ab05-40f953f09e2e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8c6703cc-4185-4dac-ab05-40f953f09e2e" (UID: "8c6703cc-4185-4dac-ab05-40f953f09e2e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.311585 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74" (UID: "6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.312005 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1e02fe72-06d2-43b5-add1-f6577e3bcc6c" (UID: "1e02fe72-06d2-43b5-add1-f6577e3bcc6c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.320607 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-kube-api-access-hqzfj" (OuterVolumeSpecName: "kube-api-access-hqzfj") pod "1e02fe72-06d2-43b5-add1-f6577e3bcc6c" (UID: "1e02fe72-06d2-43b5-add1-f6577e3bcc6c"). InnerVolumeSpecName "kube-api-access-hqzfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.321545 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-kube-api-access-fc2hq" (OuterVolumeSpecName: "kube-api-access-fc2hq") pod "6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74" (UID: "6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74"). InnerVolumeSpecName "kube-api-access-fc2hq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.324403 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c6703cc-4185-4dac-ab05-40f953f09e2e-kube-api-access-v7vbk" (OuterVolumeSpecName: "kube-api-access-v7vbk") pod "8c6703cc-4185-4dac-ab05-40f953f09e2e" (UID: "8c6703cc-4185-4dac-ab05-40f953f09e2e"). InnerVolumeSpecName "kube-api-access-v7vbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.324471 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36f77ede-a607-4d03-b939-2e8908e27def-kube-api-access-8c96j" (OuterVolumeSpecName: "kube-api-access-8c96j") pod "36f77ede-a607-4d03-b939-2e8908e27def" (UID: "36f77ede-a607-4d03-b939-2e8908e27def"). InnerVolumeSpecName "kube-api-access-8c96j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.324644 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-kube-api-access-jf8dz" (OuterVolumeSpecName: "kube-api-access-jf8dz") pod "d5b0ee6c-24bf-4b3f-875c-065c7f51862b" (UID: "d5b0ee6c-24bf-4b3f-875c-065c7f51862b"). InnerVolumeSpecName "kube-api-access-jf8dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.326658 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75288628-3722-4b4a-a76c-bd448cd5996c-kube-api-access-d9jll" (OuterVolumeSpecName: "kube-api-access-d9jll") pod "75288628-3722-4b4a-a76c-bd448cd5996c" (UID: "75288628-3722-4b4a-a76c-bd448cd5996c"). InnerVolumeSpecName "kube-api-access-d9jll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.412489 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7llg\" (UniqueName: \"kubernetes.io/projected/f32a6b81-b028-4e1d-b32b-2d8271d13875-kube-api-access-d7llg\") pod \"f32a6b81-b028-4e1d-b32b-2d8271d13875\" (UID: \"f32a6b81-b028-4e1d-b32b-2d8271d13875\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.412614 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f32a6b81-b028-4e1d-b32b-2d8271d13875-operator-scripts\") pod \"f32a6b81-b028-4e1d-b32b-2d8271d13875\" (UID: \"f32a6b81-b028-4e1d-b32b-2d8271d13875\") " Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413068 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/36f77ede-a607-4d03-b939-2e8908e27def-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413088 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf8dz\" (UniqueName: \"kubernetes.io/projected/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-kube-api-access-jf8dz\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413102 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9jll\" (UniqueName: \"kubernetes.io/projected/75288628-3722-4b4a-a76c-bd448cd5996c-kube-api-access-d9jll\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413112 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c96j\" (UniqueName: \"kubernetes.io/projected/36f77ede-a607-4d03-b939-2e8908e27def-kube-api-access-8c96j\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413121 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7vbk\" (UniqueName: \"kubernetes.io/projected/8c6703cc-4185-4dac-ab05-40f953f09e2e-kube-api-access-v7vbk\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413132 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75288628-3722-4b4a-a76c-bd448cd5996c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413140 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c6703cc-4185-4dac-ab05-40f953f09e2e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413151 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqzfj\" (UniqueName: \"kubernetes.io/projected/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-kube-api-access-hqzfj\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413159 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5b0ee6c-24bf-4b3f-875c-065c7f51862b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413167 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e02fe72-06d2-43b5-add1-f6577e3bcc6c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413177 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc2hq\" (UniqueName: \"kubernetes.io/projected/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-kube-api-access-fc2hq\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.413187 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.414207 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32a6b81-b028-4e1d-b32b-2d8271d13875-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f32a6b81-b028-4e1d-b32b-2d8271d13875" (UID: "f32a6b81-b028-4e1d-b32b-2d8271d13875"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.415490 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32a6b81-b028-4e1d-b32b-2d8271d13875-kube-api-access-d7llg" (OuterVolumeSpecName: "kube-api-access-d7llg") pod "f32a6b81-b028-4e1d-b32b-2d8271d13875" (UID: "f32a6b81-b028-4e1d-b32b-2d8271d13875"). InnerVolumeSpecName "kube-api-access-d7llg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.517504 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f32a6b81-b028-4e1d-b32b-2d8271d13875-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.517548 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7llg\" (UniqueName: \"kubernetes.io/projected/f32a6b81-b028-4e1d-b32b-2d8271d13875-kube-api-access-d7llg\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.746936 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-5d75-account-create-update-shqjl" event={"ID":"36f77ede-a607-4d03-b939-2e8908e27def","Type":"ContainerDied","Data":"3db06ca0ef6343debb1debb013b16afe6c2c0f84cf33110040750a2835a62155"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.747004 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3db06ca0ef6343debb1debb013b16afe6c2c0f84cf33110040750a2835a62155" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.747141 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5d75-account-create-update-shqjl" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.751789 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8m66z" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.752087 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8m66z" event={"ID":"6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74","Type":"ContainerDied","Data":"69e5ab2937189b3c874b8def2aa3a30b14cd315e49c1f9eac1232017104c9bec"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.752286 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69e5ab2937189b3c874b8def2aa3a30b14cd315e49c1f9eac1232017104c9bec" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.753875 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4af7-account-create-update-pmqz4" event={"ID":"75288628-3722-4b4a-a76c-bd448cd5996c","Type":"ContainerDied","Data":"55237dc6a6385a83360286012a80f7a53addae4c5074a8817a21e071fb42fc2d"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.753943 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55237dc6a6385a83360286012a80f7a53addae4c5074a8817a21e071fb42fc2d" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.754059 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4af7-account-create-update-pmqz4" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.760751 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-51e0-account-create-update-krjpx" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.760979 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-51e0-account-create-update-krjpx" event={"ID":"f32a6b81-b028-4e1d-b32b-2d8271d13875","Type":"ContainerDied","Data":"acb5de494aad3a607c058fedb05a70dbc902450a60d3824ef8e09e69d1a764fc"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.761070 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acb5de494aad3a607c058fedb05a70dbc902450a60d3824ef8e09e69d1a764fc" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.764836 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-mkljg" event={"ID":"1e02fe72-06d2-43b5-add1-f6577e3bcc6c","Type":"ContainerDied","Data":"3d9bf6f753989e154acd35e99393db6a7c853925d8697aa714ea2130218f5a11"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.764885 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d9bf6f753989e154acd35e99393db6a7c853925d8697aa714ea2130218f5a11" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.764912 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-mkljg" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.778845 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wq9lg" event={"ID":"b75d1898-5670-4f9e-88d9-69e4a1348c95","Type":"ContainerStarted","Data":"c602f8328ed1c607f7745f368d5ab19113a48a22ff9962ea7b65c3f7faf3f627"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.792146 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b62d-account-create-update-2jx8d" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.792843 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b62d-account-create-update-2jx8d" event={"ID":"d5b0ee6c-24bf-4b3f-875c-065c7f51862b","Type":"ContainerDied","Data":"012ac8c449e38b12719392aa2df33a21f6a78444ad56e2b1ba4656d6e205cadb"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.792891 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="012ac8c449e38b12719392aa2df33a21f6a78444ad56e2b1ba4656d6e205cadb" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.803732 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"2b03e85203b777515ac5854e19bc6029960e5589349057262bc2452be44aa40f"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.804017 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"8e74bdbbda8ef4828a51bb5218b7c1b5a0522a7c378847ddb5b060d81a49fea8"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.806713 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tnzh6" event={"ID":"bfcb15dd-bfcb-4519-917e-b41df5be28a9","Type":"ContainerDied","Data":"45b299bbb21d9445dd93e38cc30ec10f1fff11d6a2ec6deb5f45ac5a5edf6d2d"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.806766 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45b299bbb21d9445dd93e38cc30ec10f1fff11d6a2ec6deb5f45ac5a5edf6d2d" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.806781 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tnzh6" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.810521 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wq9lg" podStartSLOduration=4.466569769 podStartE2EDuration="9.810503823s" podCreationTimestamp="2025-12-09 17:26:43 +0000 UTC" firstStartedPulling="2025-12-09 17:26:46.660023599 +0000 UTC m=+1803.048197429" lastFinishedPulling="2025-12-09 17:26:52.003957663 +0000 UTC m=+1808.392131483" observedRunningTime="2025-12-09 17:26:52.801220453 +0000 UTC m=+1809.189394273" watchObservedRunningTime="2025-12-09 17:26:52.810503823 +0000 UTC m=+1809.198677643" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.810761 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-l48m5" event={"ID":"8c6703cc-4185-4dac-ab05-40f953f09e2e","Type":"ContainerDied","Data":"df3a3690f3c6be4e63192052e3cbcb4926065ad50428967cce32acc4710c3bcd"} Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.810866 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df3a3690f3c6be4e63192052e3cbcb4926065ad50428967cce32acc4710c3bcd" Dec 09 17:26:52 crc kubenswrapper[4954]: I1209 17:26:52.811081 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-l48m5" Dec 09 17:26:53 crc kubenswrapper[4954]: I1209 17:26:53.835048 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"a78e2220cdf96235c7f8480b72498ea076af63951db37c013bbf59d3d931b81f"} Dec 09 17:26:54 crc kubenswrapper[4954]: I1209 17:26:54.855571 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"41a7453e97c2daf007780244cfb94d474991f5752cb659c509f8479d5d3dfd4a"} Dec 09 17:26:54 crc kubenswrapper[4954]: I1209 17:26:54.855662 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"8898af52c74e97c0bc41974827d03e613d5cdd05a1653ac4e2b000d4d43d5a61"} Dec 09 17:26:54 crc kubenswrapper[4954]: I1209 17:26:54.855694 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"864730efe5b15fc52da02686f92339c085d400cba213e24789555960c11215a9"} Dec 09 17:26:55 crc kubenswrapper[4954]: I1209 17:26:55.884896 4954 generic.go:334] "Generic (PLEG): container finished" podID="8640fe28-2072-49f7-8c8d-35a2ea5432e0" containerID="eab74493a154b3368fc126aa75e9d1753a9220c6bf6879cabc7ff20617476208" exitCode=0 Dec 09 17:26:55 crc kubenswrapper[4954]: I1209 17:26:55.884986 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g4kjd" event={"ID":"8640fe28-2072-49f7-8c8d-35a2ea5432e0","Type":"ContainerDied","Data":"eab74493a154b3368fc126aa75e9d1753a9220c6bf6879cabc7ff20617476208"} Dec 09 17:26:55 crc kubenswrapper[4954]: I1209 17:26:55.900158 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"d177f7292168c4ae5236c1bf3f42af2ca8f8faf9c27d5337af5cac78cdb8b578"} Dec 09 17:26:55 crc kubenswrapper[4954]: I1209 17:26:55.900231 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"cd6fa1ac9b0ec6cd9ce89f518b4cbda0e3f2af6a6c331afc5224df195624ea13"} Dec 09 17:26:56 crc kubenswrapper[4954]: I1209 17:26:56.928938 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"2c98688f8171b669dc33d26236671e0344c7e64bd4c57d5c68029d7fd58f722b"} Dec 09 17:26:56 crc kubenswrapper[4954]: I1209 17:26:56.929350 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"159a7eb598dd8cae4e7d9609ad4c59313bb45adc77d97a0f2c9474346ec75002"} Dec 09 17:26:56 crc kubenswrapper[4954]: I1209 17:26:56.929363 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"d3fa75f7827cfa6e180d56bc163f103c3e2a2b989eb3f1e4f3369adf772f55e9"} Dec 09 17:26:56 crc kubenswrapper[4954]: I1209 17:26:56.929372 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"b7b5ed4e04fbdfe76ac7928710ec0ed0e8cfe5d4d142cf0ee36af0cf3661eb50"} Dec 09 17:26:56 crc kubenswrapper[4954]: I1209 17:26:56.929398 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69b248e7-6cb4-4805-a01f-537fe58aa42d","Type":"ContainerStarted","Data":"d07f9a86e984198f4265498d5b8e014fbec76693d90faf6def8d224e01b86a83"} Dec 09 17:26:56 crc kubenswrapper[4954]: I1209 17:26:56.933295 4954 generic.go:334] "Generic (PLEG): container finished" podID="b75d1898-5670-4f9e-88d9-69e4a1348c95" containerID="c602f8328ed1c607f7745f368d5ab19113a48a22ff9962ea7b65c3f7faf3f627" exitCode=0 Dec 09 17:26:56 crc kubenswrapper[4954]: I1209 17:26:56.933699 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wq9lg" event={"ID":"b75d1898-5670-4f9e-88d9-69e4a1348c95","Type":"ContainerDied","Data":"c602f8328ed1c607f7745f368d5ab19113a48a22ff9962ea7b65c3f7faf3f627"} Dec 09 17:26:56 crc kubenswrapper[4954]: I1209 17:26:56.974817 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=41.215777724 podStartE2EDuration="49.974784078s" podCreationTimestamp="2025-12-09 17:26:07 +0000 UTC" firstStartedPulling="2025-12-09 17:26:46.685710934 +0000 UTC m=+1803.073884754" lastFinishedPulling="2025-12-09 17:26:55.444717288 +0000 UTC m=+1811.832891108" observedRunningTime="2025-12-09 17:26:56.964883978 +0000 UTC m=+1813.353057818" watchObservedRunningTime="2025-12-09 17:26:56.974784078 +0000 UTC m=+1813.362957898" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.293145 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-f98t7"] Dec 09 17:26:57 crc kubenswrapper[4954]: E1209 17:26:57.294065 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfcb15dd-bfcb-4519-917e-b41df5be28a9" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294088 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfcb15dd-bfcb-4519-917e-b41df5be28a9" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: E1209 17:26:57.294108 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e02fe72-06d2-43b5-add1-f6577e3bcc6c" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294114 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e02fe72-06d2-43b5-add1-f6577e3bcc6c" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: E1209 17:26:57.294136 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75288628-3722-4b4a-a76c-bd448cd5996c" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294144 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="75288628-3722-4b4a-a76c-bd448cd5996c" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: E1209 17:26:57.294153 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36f77ede-a607-4d03-b939-2e8908e27def" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294161 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="36f77ede-a607-4d03-b939-2e8908e27def" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: E1209 17:26:57.294172 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294178 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: E1209 17:26:57.294188 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c6703cc-4185-4dac-ab05-40f953f09e2e" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294193 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c6703cc-4185-4dac-ab05-40f953f09e2e" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: E1209 17:26:57.294327 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5b0ee6c-24bf-4b3f-875c-065c7f51862b" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294335 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5b0ee6c-24bf-4b3f-875c-065c7f51862b" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: E1209 17:26:57.294344 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32a6b81-b028-4e1d-b32b-2d8271d13875" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294350 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32a6b81-b028-4e1d-b32b-2d8271d13875" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294551 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfcb15dd-bfcb-4519-917e-b41df5be28a9" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294564 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e02fe72-06d2-43b5-add1-f6577e3bcc6c" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294603 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c6703cc-4185-4dac-ab05-40f953f09e2e" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294615 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="75288628-3722-4b4a-a76c-bd448cd5996c" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294624 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5b0ee6c-24bf-4b3f-875c-065c7f51862b" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294635 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f32a6b81-b028-4e1d-b32b-2d8271d13875" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294644 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74" containerName="mariadb-database-create" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.294655 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="36f77ede-a607-4d03-b939-2e8908e27def" containerName="mariadb-account-create-update" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.295839 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.297870 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.306429 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-f98t7"] Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.444023 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.444207 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.444260 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsnf6\" (UniqueName: \"kubernetes.io/projected/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-kube-api-access-xsnf6\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.444304 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.444607 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-svc\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.444805 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-config\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.510558 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.546486 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-svc\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.546574 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-config\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.546657 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.546720 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.546754 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsnf6\" (UniqueName: \"kubernetes.io/projected/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-kube-api-access-xsnf6\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.546779 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.547611 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-config\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.547628 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.547820 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.548218 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-svc\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.548697 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.579030 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsnf6\" (UniqueName: \"kubernetes.io/projected/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-kube-api-access-xsnf6\") pod \"dnsmasq-dns-764c5664d7-f98t7\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.647871 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-combined-ca-bundle\") pod \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.648315 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wz8q\" (UniqueName: \"kubernetes.io/projected/8640fe28-2072-49f7-8c8d-35a2ea5432e0-kube-api-access-7wz8q\") pod \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.648364 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-config-data\") pod \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.648441 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-db-sync-config-data\") pod \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\" (UID: \"8640fe28-2072-49f7-8c8d-35a2ea5432e0\") " Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.652217 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8640fe28-2072-49f7-8c8d-35a2ea5432e0" (UID: "8640fe28-2072-49f7-8c8d-35a2ea5432e0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.653384 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8640fe28-2072-49f7-8c8d-35a2ea5432e0-kube-api-access-7wz8q" (OuterVolumeSpecName: "kube-api-access-7wz8q") pod "8640fe28-2072-49f7-8c8d-35a2ea5432e0" (UID: "8640fe28-2072-49f7-8c8d-35a2ea5432e0"). InnerVolumeSpecName "kube-api-access-7wz8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.692552 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8640fe28-2072-49f7-8c8d-35a2ea5432e0" (UID: "8640fe28-2072-49f7-8c8d-35a2ea5432e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.695791 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-config-data" (OuterVolumeSpecName: "config-data") pod "8640fe28-2072-49f7-8c8d-35a2ea5432e0" (UID: "8640fe28-2072-49f7-8c8d-35a2ea5432e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.751112 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.751161 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wz8q\" (UniqueName: \"kubernetes.io/projected/8640fe28-2072-49f7-8c8d-35a2ea5432e0-kube-api-access-7wz8q\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.751178 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.751189 4954 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8640fe28-2072-49f7-8c8d-35a2ea5432e0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:57 crc kubenswrapper[4954]: I1209 17:26:57.806722 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.008889 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g4kjd" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.009016 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g4kjd" event={"ID":"8640fe28-2072-49f7-8c8d-35a2ea5432e0","Type":"ContainerDied","Data":"d11de16b9e4b72b97a53e9c762db429bdb3bba6c9c86c62a0c31a413954fc1a4"} Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.009103 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d11de16b9e4b72b97a53e9c762db429bdb3bba6c9c86c62a0c31a413954fc1a4" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.013500 4954 generic.go:334] "Generic (PLEG): container finished" podID="e626f486-ed0d-4388-9040-1c23a41aba18" containerID="5aa994a812fd1b2f37e9c788c7399d668ca1c6caaa68ebc22a42545ef121f6d0" exitCode=0 Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.015431 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e626f486-ed0d-4388-9040-1c23a41aba18","Type":"ContainerDied","Data":"5aa994a812fd1b2f37e9c788c7399d668ca1c6caaa68ebc22a42545ef121f6d0"} Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.309070 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-f98t7"] Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.387165 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-l75lw"] Dec 09 17:26:58 crc kubenswrapper[4954]: E1209 17:26:58.387808 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8640fe28-2072-49f7-8c8d-35a2ea5432e0" containerName="glance-db-sync" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.387826 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="8640fe28-2072-49f7-8c8d-35a2ea5432e0" containerName="glance-db-sync" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.388050 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="8640fe28-2072-49f7-8c8d-35a2ea5432e0" containerName="glance-db-sync" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.389435 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.399694 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-f98t7"] Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.419497 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-l75lw"] Dec 09 17:26:58 crc kubenswrapper[4954]: W1209 17:26:58.443472 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95b76a5f_f067_4bf1_9da3_e38ff9c4b22f.slice/crio-42ebe9631e57ccca4fc13529d7294745f04db1b628681cbab5b45f388d14a45d WatchSource:0}: Error finding container 42ebe9631e57ccca4fc13529d7294745f04db1b628681cbab5b45f388d14a45d: Status 404 returned error can't find the container with id 42ebe9631e57ccca4fc13529d7294745f04db1b628681cbab5b45f388d14a45d Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.518660 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.518728 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.519100 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-config\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.519419 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.519565 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.519916 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb6w9\" (UniqueName: \"kubernetes.io/projected/8d055d0c-efc9-4476-abed-e27c50e6372a-kube-api-access-cb6w9\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.555158 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.622080 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-combined-ca-bundle\") pod \"b75d1898-5670-4f9e-88d9-69e4a1348c95\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.622247 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4zz9\" (UniqueName: \"kubernetes.io/projected/b75d1898-5670-4f9e-88d9-69e4a1348c95-kube-api-access-z4zz9\") pod \"b75d1898-5670-4f9e-88d9-69e4a1348c95\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.622273 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-config-data\") pod \"b75d1898-5670-4f9e-88d9-69e4a1348c95\" (UID: \"b75d1898-5670-4f9e-88d9-69e4a1348c95\") " Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.622803 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.622838 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.622950 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-config\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.623043 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.623102 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.623157 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb6w9\" (UniqueName: \"kubernetes.io/projected/8d055d0c-efc9-4476-abed-e27c50e6372a-kube-api-access-cb6w9\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.623762 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.625981 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.627835 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b75d1898-5670-4f9e-88d9-69e4a1348c95-kube-api-access-z4zz9" (OuterVolumeSpecName: "kube-api-access-z4zz9") pod "b75d1898-5670-4f9e-88d9-69e4a1348c95" (UID: "b75d1898-5670-4f9e-88d9-69e4a1348c95"). InnerVolumeSpecName "kube-api-access-z4zz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.635745 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.640381 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-config\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.640655 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.725935 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4zz9\" (UniqueName: \"kubernetes.io/projected/b75d1898-5670-4f9e-88d9-69e4a1348c95-kube-api-access-z4zz9\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.727182 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb6w9\" (UniqueName: \"kubernetes.io/projected/8d055d0c-efc9-4476-abed-e27c50e6372a-kube-api-access-cb6w9\") pod \"dnsmasq-dns-74f6bcbc87-l75lw\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.750867 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.753260 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b75d1898-5670-4f9e-88d9-69e4a1348c95" (UID: "b75d1898-5670-4f9e-88d9-69e4a1348c95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.765126 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-config-data" (OuterVolumeSpecName: "config-data") pod "b75d1898-5670-4f9e-88d9-69e4a1348c95" (UID: "b75d1898-5670-4f9e-88d9-69e4a1348c95"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.828627 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:58 crc kubenswrapper[4954]: I1209 17:26:58.829104 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75d1898-5670-4f9e-88d9-69e4a1348c95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.030608 4954 generic.go:334] "Generic (PLEG): container finished" podID="95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" containerID="d2447e961e29c40cfb67b87da624bead701206a6aafc7a9ce04c9c80c8c23207" exitCode=0 Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.030739 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-f98t7" event={"ID":"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f","Type":"ContainerDied","Data":"d2447e961e29c40cfb67b87da624bead701206a6aafc7a9ce04c9c80c8c23207"} Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.030781 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-f98t7" event={"ID":"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f","Type":"ContainerStarted","Data":"42ebe9631e57ccca4fc13529d7294745f04db1b628681cbab5b45f388d14a45d"} Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.036941 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e626f486-ed0d-4388-9040-1c23a41aba18","Type":"ContainerStarted","Data":"6093515770e90814d488365df2b67e489254c9db4ca511f27cbd8387eebf0c6c"} Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.039613 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wq9lg" event={"ID":"b75d1898-5670-4f9e-88d9-69e4a1348c95","Type":"ContainerDied","Data":"632ef31dec5aea84c1a4c71791de6c946aec0269c84229f41c3cb2b16e862beb"} Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.039648 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="632ef31dec5aea84c1a4c71791de6c946aec0269c84229f41c3cb2b16e862beb" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.039694 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wq9lg" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.254983 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-l75lw"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.283099 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-l75lw"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.311628 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4l5k6"] Dec 09 17:26:59 crc kubenswrapper[4954]: E1209 17:26:59.312149 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b75d1898-5670-4f9e-88d9-69e4a1348c95" containerName="keystone-db-sync" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.312175 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b75d1898-5670-4f9e-88d9-69e4a1348c95" containerName="keystone-db-sync" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.312468 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="b75d1898-5670-4f9e-88d9-69e4a1348c95" containerName="keystone-db-sync" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.344103 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.364728 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ddwb9"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.366674 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.368897 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.371644 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ddwb9"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.376982 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.377303 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.377459 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.377684 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5xlkl" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.385138 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4l5k6"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.414648 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-p5ndh"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.416094 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.419761 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-xcmzr" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.420261 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.446483 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-p5ndh"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462567 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462657 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-scripts\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462691 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-combined-ca-bundle\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462717 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkrqv\" (UniqueName: \"kubernetes.io/projected/53d056c1-e346-4c89-921d-43f5827d2993-kube-api-access-zkrqv\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462774 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-config\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462803 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-credential-keys\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462821 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wp75\" (UniqueName: \"kubernetes.io/projected/25667d18-6643-4344-9ac3-e3a4f1af6957-kube-api-access-5wp75\") pod \"heat-db-sync-p5ndh\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462858 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-config-data\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462898 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-combined-ca-bundle\") pod \"heat-db-sync-p5ndh\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462920 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-svc\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462941 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-fernet-keys\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462958 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.462979 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-config-data\") pod \"heat-db-sync-p5ndh\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.463027 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwzwh\" (UniqueName: \"kubernetes.io/projected/bc313343-7599-4dad-bda6-ecdb04007e4f-kube-api-access-fwzwh\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.463049 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.606816 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-cqb6p"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.607786 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-config\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.607863 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-credential-keys\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.607892 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wp75\" (UniqueName: \"kubernetes.io/projected/25667d18-6643-4344-9ac3-e3a4f1af6957-kube-api-access-5wp75\") pod \"heat-db-sync-p5ndh\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.607981 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-config-data\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608067 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-combined-ca-bundle\") pod \"heat-db-sync-p5ndh\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608111 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-svc\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608145 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-fernet-keys\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608179 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608219 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-config-data\") pod \"heat-db-sync-p5ndh\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608274 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwzwh\" (UniqueName: \"kubernetes.io/projected/bc313343-7599-4dad-bda6-ecdb04007e4f-kube-api-access-fwzwh\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608306 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608447 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608536 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-scripts\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608613 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-combined-ca-bundle\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.608662 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkrqv\" (UniqueName: \"kubernetes.io/projected/53d056c1-e346-4c89-921d-43f5827d2993-kube-api-access-zkrqv\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.609305 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.614248 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.614943 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.617287 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-fernet-keys\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.618158 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-config\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.625446 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-config-data\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.626108 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-svc\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.627112 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-credential-keys\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.627622 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.632271 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-combined-ca-bundle\") pod \"heat-db-sync-p5ndh\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.632811 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fd8bh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.634636 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.635261 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.687853 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-config-data\") pod \"heat-db-sync-p5ndh\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.691464 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-combined-ca-bundle\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.692012 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-scripts\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.700692 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwzwh\" (UniqueName: \"kubernetes.io/projected/bc313343-7599-4dad-bda6-ecdb04007e4f-kube-api-access-fwzwh\") pod \"dnsmasq-dns-847c4cc679-4l5k6\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.701174 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkrqv\" (UniqueName: \"kubernetes.io/projected/53d056c1-e346-4c89-921d-43f5827d2993-kube-api-access-zkrqv\") pod \"keystone-bootstrap-ddwb9\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.704331 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wp75\" (UniqueName: \"kubernetes.io/projected/25667d18-6643-4344-9ac3-e3a4f1af6957-kube-api-access-5wp75\") pod \"heat-db-sync-p5ndh\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.713057 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.716050 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-config\") pod \"neutron-db-sync-cqb6p\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.716314 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-combined-ca-bundle\") pod \"neutron-db-sync-cqb6p\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.716464 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcnzf\" (UniqueName: \"kubernetes.io/projected/4296b9c8-795b-4786-80a8-1b8c8d723451-kube-api-access-xcnzf\") pod \"neutron-db-sync-cqb6p\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.734820 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-cqb6p"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.773126 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.783497 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-p5ndh" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.815349 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-fsxd4"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.817398 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fsxd4" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.825113 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcnzf\" (UniqueName: \"kubernetes.io/projected/4296b9c8-795b-4786-80a8-1b8c8d723451-kube-api-access-xcnzf\") pod \"neutron-db-sync-cqb6p\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.825296 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-config\") pod \"neutron-db-sync-cqb6p\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.825384 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-combined-ca-bundle\") pod \"neutron-db-sync-cqb6p\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.833078 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.833175 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.833221 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-4ztm9" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.853579 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-combined-ca-bundle\") pod \"neutron-db-sync-cqb6p\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.869820 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-fsxd4"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.871294 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-config\") pod \"neutron-db-sync-cqb6p\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.875816 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcnzf\" (UniqueName: \"kubernetes.io/projected/4296b9c8-795b-4786-80a8-1b8c8d723451-kube-api-access-xcnzf\") pod \"neutron-db-sync-cqb6p\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.918947 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-c686v"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.920868 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c686v" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.928177 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fd76337-c1f2-43ff-b021-8837dd046dc2-logs\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.928233 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-scripts\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.928375 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-config-data\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.928410 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp2f2\" (UniqueName: \"kubernetes.io/projected/5fd76337-c1f2-43ff-b021-8837dd046dc2-kube-api-access-wp2f2\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.928526 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-combined-ca-bundle\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.928900 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-n4cmp" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.941494 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.951677 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-c686v"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.968665 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-r98b8"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.970527 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r98b8" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.973917 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-r98b8"] Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.980633 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 09 17:26:59 crc kubenswrapper[4954]: I1209 17:26:59.981072 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.004075 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-npkrm" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.007442 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4l5k6"] Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.021673 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-77x96"] Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.024375 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.030965 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.041091 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp2f2\" (UniqueName: \"kubernetes.io/projected/5fd76337-c1f2-43ff-b021-8837dd046dc2-kube-api-access-wp2f2\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.041263 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq92f\" (UniqueName: \"kubernetes.io/projected/3b087316-6f76-4aaa-bd43-d5b01051d4b2-kube-api-access-tq92f\") pod \"barbican-db-sync-c686v\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.041354 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-combined-ca-bundle\") pod \"barbican-db-sync-c686v\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.041409 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-combined-ca-bundle\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.041568 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fd76337-c1f2-43ff-b021-8837dd046dc2-logs\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.041628 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-scripts\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.041730 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-db-sync-config-data\") pod \"barbican-db-sync-c686v\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.041857 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-config-data\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.063221 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-77x96"] Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.071100 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fd76337-c1f2-43ff-b021-8837dd046dc2-logs\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.099561 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-combined-ca-bundle\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.106723 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-config-data\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.117823 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.139882 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-scripts\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.146997 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.147116 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4kr6\" (UniqueName: \"kubernetes.io/projected/9653f9d1-0297-458e-bfb6-b6801ef96c3c-kube-api-access-l4kr6\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.147168 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.147216 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-db-sync-config-data\") pod \"barbican-db-sync-c686v\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.147374 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.147405 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-config-data\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.147428 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mvwq\" (UniqueName: \"kubernetes.io/projected/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-kube-api-access-9mvwq\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.147500 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-combined-ca-bundle\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.147546 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq92f\" (UniqueName: \"kubernetes.io/projected/3b087316-6f76-4aaa-bd43-d5b01051d4b2-kube-api-access-tq92f\") pod \"barbican-db-sync-c686v\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.147562 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-db-sync-config-data\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.147613 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9653f9d1-0297-458e-bfb6-b6801ef96c3c-etc-machine-id\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.148441 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-scripts\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.148483 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-config\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.148539 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-combined-ca-bundle\") pod \"barbican-db-sync-c686v\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.148559 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.149016 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp2f2\" (UniqueName: \"kubernetes.io/projected/5fd76337-c1f2-43ff-b021-8837dd046dc2-kube-api-access-wp2f2\") pod \"placement-db-sync-fsxd4\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.166613 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-combined-ca-bundle\") pod \"barbican-db-sync-c686v\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.174786 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-f98t7" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.219964 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-db-sync-config-data\") pod \"barbican-db-sync-c686v\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.220041 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq92f\" (UniqueName: \"kubernetes.io/projected/3b087316-6f76-4aaa-bd43-d5b01051d4b2-kube-api-access-tq92f\") pod \"barbican-db-sync-c686v\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.241298 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" event={"ID":"8d055d0c-efc9-4476-abed-e27c50e6372a","Type":"ContainerStarted","Data":"2ab2d7574cf8741e95e73dbb601e5e5c6e389a535d0b2dd28258ef0e0f6c9d6b"} Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.241377 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-f98t7" event={"ID":"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f","Type":"ContainerDied","Data":"42ebe9631e57ccca4fc13529d7294745f04db1b628681cbab5b45f388d14a45d"} Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.241450 4954 scope.go:117] "RemoveContainer" containerID="d2447e961e29c40cfb67b87da624bead701206a6aafc7a9ce04c9c80c8c23207" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.249821 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-config\") pod \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.249908 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-svc\") pod \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.249955 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-swift-storage-0\") pod \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.250043 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-nb\") pod \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.250067 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsnf6\" (UniqueName: \"kubernetes.io/projected/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-kube-api-access-xsnf6\") pod \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.250103 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-sb\") pod \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\" (UID: \"95b76a5f-f067-4bf1-9da3-e38ff9c4b22f\") " Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.250453 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.250485 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-config-data\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.250503 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mvwq\" (UniqueName: \"kubernetes.io/projected/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-kube-api-access-9mvwq\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.250535 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-combined-ca-bundle\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.250565 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-db-sync-config-data\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.252287 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9653f9d1-0297-458e-bfb6-b6801ef96c3c-etc-machine-id\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.252323 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-scripts\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.252355 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-config\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.252399 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.252530 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.252586 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4kr6\" (UniqueName: \"kubernetes.io/projected/9653f9d1-0297-458e-bfb6-b6801ef96c3c-kube-api-access-l4kr6\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.252645 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.253539 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.255679 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.267243 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.268178 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-config\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.268850 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9653f9d1-0297-458e-bfb6-b6801ef96c3c-etc-machine-id\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.269413 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-scripts\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.270265 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.272688 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-combined-ca-bundle\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.292804 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mvwq\" (UniqueName: \"kubernetes.io/projected/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-kube-api-access-9mvwq\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.293740 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-config-data\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.295885 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-77x96\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.296148 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-db-sync-config-data\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.299751 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.324190 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4kr6\" (UniqueName: \"kubernetes.io/projected/9653f9d1-0297-458e-bfb6-b6801ef96c3c-kube-api-access-l4kr6\") pod \"cinder-db-sync-r98b8\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.363954 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-kube-api-access-xsnf6" (OuterVolumeSpecName: "kube-api-access-xsnf6") pod "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" (UID: "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f"). InnerVolumeSpecName "kube-api-access-xsnf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.367888 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsnf6\" (UniqueName: \"kubernetes.io/projected/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-kube-api-access-xsnf6\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.386316 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.421672 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.460938 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" (UID: "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.469850 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.502454 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:27:00 crc kubenswrapper[4954]: E1209 17:27:00.503099 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" containerName="init" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.503119 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" containerName="init" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.503419 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" containerName="init" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.506246 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.513851 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.514470 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 17:27:00 crc kubenswrapper[4954]: I1209 17:27:00.517569 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.744237 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd225\" (UniqueName: \"kubernetes.io/projected/565df52a-04b4-463b-a1b1-59eee6d94ea5-kube-api-access-vd225\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.744709 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-scripts\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.744747 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-log-httpd\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.744832 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.744904 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-config-data\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.744928 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-run-httpd\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.744951 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.830391 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4l5k6"] Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.859223 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd225\" (UniqueName: \"kubernetes.io/projected/565df52a-04b4-463b-a1b1-59eee6d94ea5-kube-api-access-vd225\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.859295 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-scripts\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.859334 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-log-httpd\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.859412 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.863795 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-config-data\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.864228 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-run-httpd\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.864263 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.879582 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-log-httpd\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.879902 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-run-httpd\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.945298 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.948555 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-scripts\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.955793 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd225\" (UniqueName: \"kubernetes.io/projected/565df52a-04b4-463b-a1b1-59eee6d94ea5-kube-api-access-vd225\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.956074 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-config-data\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:01 crc kubenswrapper[4954]: I1209 17:27:01.956372 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " pod="openstack/ceilometer-0" Dec 09 17:27:02 crc kubenswrapper[4954]: W1209 17:27:02.053021 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc313343_7599_4dad_bda6_ecdb04007e4f.slice/crio-927c78ed3176673f8ee271316ab9440cfd28b46d154696d8816f2ea7625c8253 WatchSource:0}: Error finding container 927c78ed3176673f8ee271316ab9440cfd28b46d154696d8816f2ea7625c8253: Status 404 returned error can't find the container with id 927c78ed3176673f8ee271316ab9440cfd28b46d154696d8816f2ea7625c8253 Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.291703 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-p5ndh"] Dec 09 17:27:02 crc kubenswrapper[4954]: W1209 17:27:02.313280 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25667d18_6643_4344_9ac3_e3a4f1af6957.slice/crio-ea4239307ccedc6fc923097af10c147bf617eb10495c775c3dda6207306a296c WatchSource:0}: Error finding container ea4239307ccedc6fc923097af10c147bf617eb10495c775c3dda6207306a296c: Status 404 returned error can't find the container with id ea4239307ccedc6fc923097af10c147bf617eb10495c775c3dda6207306a296c Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.379993 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.572647 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ddwb9"] Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.649702 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-cqb6p"] Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.716543 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-fsxd4"] Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.751264 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-c686v"] Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.762175 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-r98b8"] Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.831957 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-p5ndh" event={"ID":"25667d18-6643-4344-9ac3-e3a4f1af6957","Type":"ContainerStarted","Data":"ea4239307ccedc6fc923097af10c147bf617eb10495c775c3dda6207306a296c"} Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.837223 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r98b8" event={"ID":"9653f9d1-0297-458e-bfb6-b6801ef96c3c","Type":"ContainerStarted","Data":"7e5d8deda681bc17d1ebb32ba6ecf702f9d9a41da2de897794ea7fb2165a106d"} Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.842298 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cqb6p" event={"ID":"4296b9c8-795b-4786-80a8-1b8c8d723451","Type":"ContainerStarted","Data":"771f2ca1e031f1cb318bea0fee93a7aa26646a4c4e2c814b8b80635ddc42182f"} Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.864322 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" event={"ID":"bc313343-7599-4dad-bda6-ecdb04007e4f","Type":"ContainerStarted","Data":"8554536c95787bb2f87dca7082e239161ec0e5e58bddfc5fbea71c405187085a"} Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.864390 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" event={"ID":"bc313343-7599-4dad-bda6-ecdb04007e4f","Type":"ContainerStarted","Data":"927c78ed3176673f8ee271316ab9440cfd28b46d154696d8816f2ea7625c8253"} Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.864731 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" podUID="bc313343-7599-4dad-bda6-ecdb04007e4f" containerName="init" containerID="cri-o://8554536c95787bb2f87dca7082e239161ec0e5e58bddfc5fbea71c405187085a" gracePeriod=10 Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.866368 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ddwb9" event={"ID":"53d056c1-e346-4c89-921d-43f5827d2993","Type":"ContainerStarted","Data":"3bf2c9558830502b6e66eed02463a878691e1aa93251e7189e6f20aff0662c11"} Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.870616 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e626f486-ed0d-4388-9040-1c23a41aba18","Type":"ContainerStarted","Data":"57645460d8342ff081db6136d4ee0a2fd3058e321014fd404d3a29c0eade6f31"} Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.872432 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c686v" event={"ID":"3b087316-6f76-4aaa-bd43-d5b01051d4b2","Type":"ContainerStarted","Data":"fb492d5eafcaef5a0ec12ac9ebc9ccf2bdec9a40db78134e0c7c7e54bf514204"} Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.873952 4954 generic.go:334] "Generic (PLEG): container finished" podID="8d055d0c-efc9-4476-abed-e27c50e6372a" containerID="b01cdc918c27975f6b87845044daf3ec60a8c0b031f5b0ffe36dc99affeda72a" exitCode=0 Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.873994 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" event={"ID":"8d055d0c-efc9-4476-abed-e27c50e6372a","Type":"ContainerDied","Data":"b01cdc918c27975f6b87845044daf3ec60a8c0b031f5b0ffe36dc99affeda72a"} Dec 09 17:27:02 crc kubenswrapper[4954]: I1209 17:27:02.881443 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fsxd4" event={"ID":"5fd76337-c1f2-43ff-b021-8837dd046dc2","Type":"ContainerStarted","Data":"d2eeaa2edb192357d38d5aef2930c8b5dc195ec06ca87a6975ce9d748377c000"} Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.137582 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.154079 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-77x96"] Dec 09 17:27:03 crc kubenswrapper[4954]: W1209 17:27:03.167845 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod565df52a_04b4_463b_a1b1_59eee6d94ea5.slice/crio-68ca007837b4e705a4238cfe1f271e417faacdbf93bf0f4fcdc0b934207f2560 WatchSource:0}: Error finding container 68ca007837b4e705a4238cfe1f271e417faacdbf93bf0f4fcdc0b934207f2560: Status 404 returned error can't find the container with id 68ca007837b4e705a4238cfe1f271e417faacdbf93bf0f4fcdc0b934207f2560 Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.430716 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" (UID: "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.431093 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-config" (OuterVolumeSpecName: "config") pod "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" (UID: "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.431268 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" (UID: "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.431907 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" (UID: "95b76a5f-f067-4bf1-9da3-e38ff9c4b22f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.493917 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.493951 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.493962 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.493972 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.564334 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.637285 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-f98t7"] Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.643671 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-f98t7"] Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.696572 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb6w9\" (UniqueName: \"kubernetes.io/projected/8d055d0c-efc9-4476-abed-e27c50e6372a-kube-api-access-cb6w9\") pod \"8d055d0c-efc9-4476-abed-e27c50e6372a\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.696769 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-sb\") pod \"8d055d0c-efc9-4476-abed-e27c50e6372a\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.696907 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-svc\") pod \"8d055d0c-efc9-4476-abed-e27c50e6372a\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.696928 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-swift-storage-0\") pod \"8d055d0c-efc9-4476-abed-e27c50e6372a\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.696952 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-nb\") pod \"8d055d0c-efc9-4476-abed-e27c50e6372a\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.696973 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-config\") pod \"8d055d0c-efc9-4476-abed-e27c50e6372a\" (UID: \"8d055d0c-efc9-4476-abed-e27c50e6372a\") " Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.715359 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d055d0c-efc9-4476-abed-e27c50e6372a-kube-api-access-cb6w9" (OuterVolumeSpecName: "kube-api-access-cb6w9") pod "8d055d0c-efc9-4476-abed-e27c50e6372a" (UID: "8d055d0c-efc9-4476-abed-e27c50e6372a"). InnerVolumeSpecName "kube-api-access-cb6w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.740982 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8d055d0c-efc9-4476-abed-e27c50e6372a" (UID: "8d055d0c-efc9-4476-abed-e27c50e6372a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.751420 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8d055d0c-efc9-4476-abed-e27c50e6372a" (UID: "8d055d0c-efc9-4476-abed-e27c50e6372a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.758393 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-config" (OuterVolumeSpecName: "config") pod "8d055d0c-efc9-4476-abed-e27c50e6372a" (UID: "8d055d0c-efc9-4476-abed-e27c50e6372a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.777435 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8d055d0c-efc9-4476-abed-e27c50e6372a" (UID: "8d055d0c-efc9-4476-abed-e27c50e6372a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.790015 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8d055d0c-efc9-4476-abed-e27c50e6372a" (UID: "8d055d0c-efc9-4476-abed-e27c50e6372a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.811653 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.811703 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.811718 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.811731 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.811742 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb6w9\" (UniqueName: \"kubernetes.io/projected/8d055d0c-efc9-4476-abed-e27c50e6372a-kube-api-access-cb6w9\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.811753 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d055d0c-efc9-4476-abed-e27c50e6372a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.902487 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ddwb9" event={"ID":"53d056c1-e346-4c89-921d-43f5827d2993","Type":"ContainerStarted","Data":"ce673116279a476e69c8b4fbe524f61413724f7e556919489ae5879984bf82d7"} Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.906130 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e626f486-ed0d-4388-9040-1c23a41aba18","Type":"ContainerStarted","Data":"2546016937d3bb6588a7b70a912ee2771bdb8558d3e5ff4837994c7a66edd5ed"} Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.908344 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" event={"ID":"83aa4b70-c6f9-413f-9d19-65f2b9701ecf","Type":"ContainerStarted","Data":"327b98bd762c21f81bf271a885bea9eccb06aa0d15565894ecf17d60f11b6d50"} Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.909492 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"565df52a-04b4-463b-a1b1-59eee6d94ea5","Type":"ContainerStarted","Data":"68ca007837b4e705a4238cfe1f271e417faacdbf93bf0f4fcdc0b934207f2560"} Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.911902 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cqb6p" event={"ID":"4296b9c8-795b-4786-80a8-1b8c8d723451","Type":"ContainerStarted","Data":"4104c8ec76bad3c56a6c514d99d167e0bb6322b543aa3dbff81c00a9d841ec15"} Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.916178 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.916125 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-l75lw" event={"ID":"8d055d0c-efc9-4476-abed-e27c50e6372a","Type":"ContainerDied","Data":"2ab2d7574cf8741e95e73dbb601e5e5c6e389a535d0b2dd28258ef0e0f6c9d6b"} Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.919495 4954 scope.go:117] "RemoveContainer" containerID="b01cdc918c27975f6b87845044daf3ec60a8c0b031f5b0ffe36dc99affeda72a" Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.920909 4954 generic.go:334] "Generic (PLEG): container finished" podID="bc313343-7599-4dad-bda6-ecdb04007e4f" containerID="8554536c95787bb2f87dca7082e239161ec0e5e58bddfc5fbea71c405187085a" exitCode=0 Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.920979 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" event={"ID":"bc313343-7599-4dad-bda6-ecdb04007e4f","Type":"ContainerDied","Data":"8554536c95787bb2f87dca7082e239161ec0e5e58bddfc5fbea71c405187085a"} Dec 09 17:27:03 crc kubenswrapper[4954]: I1209 17:27:03.960640 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.960611293 podStartE2EDuration="18.960611293s" podCreationTimestamp="2025-12-09 17:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:03.943713344 +0000 UTC m=+1820.331887164" watchObservedRunningTime="2025-12-09 17:27:03.960611293 +0000 UTC m=+1820.348785113" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.144124 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-cqb6p" podStartSLOduration=5.144094686 podStartE2EDuration="5.144094686s" podCreationTimestamp="2025-12-09 17:26:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:03.997128785 +0000 UTC m=+1820.385302605" watchObservedRunningTime="2025-12-09 17:27:04.144094686 +0000 UTC m=+1820.532268506" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.220515 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95b76a5f-f067-4bf1-9da3-e38ff9c4b22f" path="/var/lib/kubelet/pods/95b76a5f-f067-4bf1-9da3-e38ff9c4b22f/volumes" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.223301 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.226694 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:27:04 crc kubenswrapper[4954]: E1209 17:27:04.227516 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.235401 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-l75lw"] Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.257004 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-l75lw"] Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.299166 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.407204 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-config\") pod \"bc313343-7599-4dad-bda6-ecdb04007e4f\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.407392 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-sb\") pod \"bc313343-7599-4dad-bda6-ecdb04007e4f\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.407450 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-nb\") pod \"bc313343-7599-4dad-bda6-ecdb04007e4f\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.407498 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwzwh\" (UniqueName: \"kubernetes.io/projected/bc313343-7599-4dad-bda6-ecdb04007e4f-kube-api-access-fwzwh\") pod \"bc313343-7599-4dad-bda6-ecdb04007e4f\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.407529 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-svc\") pod \"bc313343-7599-4dad-bda6-ecdb04007e4f\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.407622 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-swift-storage-0\") pod \"bc313343-7599-4dad-bda6-ecdb04007e4f\" (UID: \"bc313343-7599-4dad-bda6-ecdb04007e4f\") " Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.415501 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc313343-7599-4dad-bda6-ecdb04007e4f-kube-api-access-fwzwh" (OuterVolumeSpecName: "kube-api-access-fwzwh") pod "bc313343-7599-4dad-bda6-ecdb04007e4f" (UID: "bc313343-7599-4dad-bda6-ecdb04007e4f"). InnerVolumeSpecName "kube-api-access-fwzwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.442111 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bc313343-7599-4dad-bda6-ecdb04007e4f" (UID: "bc313343-7599-4dad-bda6-ecdb04007e4f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.453874 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-config" (OuterVolumeSpecName: "config") pod "bc313343-7599-4dad-bda6-ecdb04007e4f" (UID: "bc313343-7599-4dad-bda6-ecdb04007e4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.454992 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bc313343-7599-4dad-bda6-ecdb04007e4f" (UID: "bc313343-7599-4dad-bda6-ecdb04007e4f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.460524 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bc313343-7599-4dad-bda6-ecdb04007e4f" (UID: "bc313343-7599-4dad-bda6-ecdb04007e4f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.478942 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bc313343-7599-4dad-bda6-ecdb04007e4f" (UID: "bc313343-7599-4dad-bda6-ecdb04007e4f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.512048 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.512087 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.512103 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.512115 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.512130 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwzwh\" (UniqueName: \"kubernetes.io/projected/bc313343-7599-4dad-bda6-ecdb04007e4f-kube-api-access-fwzwh\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.512140 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc313343-7599-4dad-bda6-ecdb04007e4f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.943817 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.943837 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-4l5k6" event={"ID":"bc313343-7599-4dad-bda6-ecdb04007e4f","Type":"ContainerDied","Data":"927c78ed3176673f8ee271316ab9440cfd28b46d154696d8816f2ea7625c8253"} Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.944911 4954 scope.go:117] "RemoveContainer" containerID="8554536c95787bb2f87dca7082e239161ec0e5e58bddfc5fbea71c405187085a" Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.958079 4954 generic.go:334] "Generic (PLEG): container finished" podID="83aa4b70-c6f9-413f-9d19-65f2b9701ecf" containerID="a8e4b24d4b0576a1066c670e11eaaaff03001241fb677f63cdc51997f76a1560" exitCode=0 Dec 09 17:27:04 crc kubenswrapper[4954]: I1209 17:27:04.958304 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" event={"ID":"83aa4b70-c6f9-413f-9d19-65f2b9701ecf","Type":"ContainerDied","Data":"a8e4b24d4b0576a1066c670e11eaaaff03001241fb677f63cdc51997f76a1560"} Dec 09 17:27:05 crc kubenswrapper[4954]: I1209 17:27:05.005608 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ddwb9" podStartSLOduration=6.005570715 podStartE2EDuration="6.005570715s" podCreationTimestamp="2025-12-09 17:26:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:04.984843747 +0000 UTC m=+1821.373017577" watchObservedRunningTime="2025-12-09 17:27:05.005570715 +0000 UTC m=+1821.393744535" Dec 09 17:27:05 crc kubenswrapper[4954]: I1209 17:27:05.125689 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4l5k6"] Dec 09 17:27:05 crc kubenswrapper[4954]: I1209 17:27:05.157394 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-4l5k6"] Dec 09 17:27:05 crc kubenswrapper[4954]: I1209 17:27:05.974930 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" event={"ID":"83aa4b70-c6f9-413f-9d19-65f2b9701ecf","Type":"ContainerStarted","Data":"aaaa514a40281435d4c5223f2f03d3f099b06add2067aa09d4d3b9640768888f"} Dec 09 17:27:05 crc kubenswrapper[4954]: I1209 17:27:05.975254 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:06 crc kubenswrapper[4954]: I1209 17:27:06.003840 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 09 17:27:06 crc kubenswrapper[4954]: I1209 17:27:06.013227 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" podStartSLOduration=7.013206729 podStartE2EDuration="7.013206729s" podCreationTimestamp="2025-12-09 17:26:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:06.003766054 +0000 UTC m=+1822.391939874" watchObservedRunningTime="2025-12-09 17:27:06.013206729 +0000 UTC m=+1822.401380549" Dec 09 17:27:06 crc kubenswrapper[4954]: I1209 17:27:06.137514 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d055d0c-efc9-4476-abed-e27c50e6372a" path="/var/lib/kubelet/pods/8d055d0c-efc9-4476-abed-e27c50e6372a/volumes" Dec 09 17:27:06 crc kubenswrapper[4954]: I1209 17:27:06.138448 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc313343-7599-4dad-bda6-ecdb04007e4f" path="/var/lib/kubelet/pods/bc313343-7599-4dad-bda6-ecdb04007e4f/volumes" Dec 09 17:27:09 crc kubenswrapper[4954]: I1209 17:27:09.016005 4954 generic.go:334] "Generic (PLEG): container finished" podID="53d056c1-e346-4c89-921d-43f5827d2993" containerID="ce673116279a476e69c8b4fbe524f61413724f7e556919489ae5879984bf82d7" exitCode=0 Dec 09 17:27:09 crc kubenswrapper[4954]: I1209 17:27:09.016689 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ddwb9" event={"ID":"53d056c1-e346-4c89-921d-43f5827d2993","Type":"ContainerDied","Data":"ce673116279a476e69c8b4fbe524f61413724f7e556919489ae5879984bf82d7"} Dec 09 17:27:10 crc kubenswrapper[4954]: I1209 17:27:10.424358 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:10 crc kubenswrapper[4954]: I1209 17:27:10.512437 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-q85b8"] Dec 09 17:27:10 crc kubenswrapper[4954]: I1209 17:27:10.513232 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-q85b8" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerName="dnsmasq-dns" containerID="cri-o://36b7c152baf598616164b052b73516434a385be71cac9598195a5390df8eb505" gracePeriod=10 Dec 09 17:27:11 crc kubenswrapper[4954]: I1209 17:27:11.044338 4954 generic.go:334] "Generic (PLEG): container finished" podID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerID="36b7c152baf598616164b052b73516434a385be71cac9598195a5390df8eb505" exitCode=0 Dec 09 17:27:11 crc kubenswrapper[4954]: I1209 17:27:11.044421 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-q85b8" event={"ID":"7801abe0-4c3f-4251-8b29-e8ff3e11b580","Type":"ContainerDied","Data":"36b7c152baf598616164b052b73516434a385be71cac9598195a5390df8eb505"} Dec 09 17:27:13 crc kubenswrapper[4954]: I1209 17:27:13.050143 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-q85b8" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: connect: connection refused" Dec 09 17:27:16 crc kubenswrapper[4954]: I1209 17:27:16.002840 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 09 17:27:16 crc kubenswrapper[4954]: I1209 17:27:16.010508 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 09 17:27:16 crc kubenswrapper[4954]: I1209 17:27:16.115178 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 09 17:27:17 crc kubenswrapper[4954]: I1209 17:27:17.120755 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:27:17 crc kubenswrapper[4954]: E1209 17:27:17.121129 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:27:17 crc kubenswrapper[4954]: E1209 17:27:17.122168 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 09 17:27:17 crc kubenswrapper[4954]: E1209 17:27:17.122361 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tq92f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-c686v_openstack(3b087316-6f76-4aaa-bd43-d5b01051d4b2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:27:17 crc kubenswrapper[4954]: E1209 17:27:17.123564 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-c686v" podUID="3b087316-6f76-4aaa-bd43-d5b01051d4b2" Dec 09 17:27:18 crc kubenswrapper[4954]: I1209 17:27:18.051017 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-q85b8" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: connect: connection refused" Dec 09 17:27:18 crc kubenswrapper[4954]: E1209 17:27:18.139631 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-c686v" podUID="3b087316-6f76-4aaa-bd43-d5b01051d4b2" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.170064 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ddwb9" event={"ID":"53d056c1-e346-4c89-921d-43f5827d2993","Type":"ContainerDied","Data":"3bf2c9558830502b6e66eed02463a878691e1aa93251e7189e6f20aff0662c11"} Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.170524 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bf2c9558830502b6e66eed02463a878691e1aa93251e7189e6f20aff0662c11" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.198250 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.276402 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-fernet-keys\") pod \"53d056c1-e346-4c89-921d-43f5827d2993\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.276458 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-credential-keys\") pod \"53d056c1-e346-4c89-921d-43f5827d2993\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.276516 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-scripts\") pod \"53d056c1-e346-4c89-921d-43f5827d2993\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.276692 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkrqv\" (UniqueName: \"kubernetes.io/projected/53d056c1-e346-4c89-921d-43f5827d2993-kube-api-access-zkrqv\") pod \"53d056c1-e346-4c89-921d-43f5827d2993\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.276727 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-combined-ca-bundle\") pod \"53d056c1-e346-4c89-921d-43f5827d2993\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.276891 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-config-data\") pod \"53d056c1-e346-4c89-921d-43f5827d2993\" (UID: \"53d056c1-e346-4c89-921d-43f5827d2993\") " Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.294805 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-scripts" (OuterVolumeSpecName: "scripts") pod "53d056c1-e346-4c89-921d-43f5827d2993" (UID: "53d056c1-e346-4c89-921d-43f5827d2993"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.294755 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "53d056c1-e346-4c89-921d-43f5827d2993" (UID: "53d056c1-e346-4c89-921d-43f5827d2993"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.294889 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "53d056c1-e346-4c89-921d-43f5827d2993" (UID: "53d056c1-e346-4c89-921d-43f5827d2993"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.295513 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d056c1-e346-4c89-921d-43f5827d2993-kube-api-access-zkrqv" (OuterVolumeSpecName: "kube-api-access-zkrqv") pod "53d056c1-e346-4c89-921d-43f5827d2993" (UID: "53d056c1-e346-4c89-921d-43f5827d2993"). InnerVolumeSpecName "kube-api-access-zkrqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.307074 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53d056c1-e346-4c89-921d-43f5827d2993" (UID: "53d056c1-e346-4c89-921d-43f5827d2993"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.307496 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-config-data" (OuterVolumeSpecName: "config-data") pod "53d056c1-e346-4c89-921d-43f5827d2993" (UID: "53d056c1-e346-4c89-921d-43f5827d2993"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.386290 4954 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.386411 4954 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.386428 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.386440 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkrqv\" (UniqueName: \"kubernetes.io/projected/53d056c1-e346-4c89-921d-43f5827d2993-kube-api-access-zkrqv\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.386451 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:21 crc kubenswrapper[4954]: I1209 17:27:21.386461 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d056c1-e346-4c89-921d-43f5827d2993-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.184797 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ddwb9" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.334741 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ddwb9"] Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.346857 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ddwb9"] Dec 09 17:27:22 crc kubenswrapper[4954]: E1209 17:27:22.430790 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 09 17:27:22 crc kubenswrapper[4954]: E1209 17:27:22.431068 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n54ch5fbh59h5c4h548h55ch58dh599h5d6h5c8h556h67fh54fh559h55ch8fh77h5f4hfchffh96h649h5bch555h599h9bh575h685h569h688hf9h64q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vd225,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(565df52a-04b4-463b-a1b1-59eee6d94ea5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.481142 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-cb6lc"] Dec 09 17:27:22 crc kubenswrapper[4954]: E1209 17:27:22.481883 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc313343-7599-4dad-bda6-ecdb04007e4f" containerName="init" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.481912 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc313343-7599-4dad-bda6-ecdb04007e4f" containerName="init" Dec 09 17:27:22 crc kubenswrapper[4954]: E1209 17:27:22.481950 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d056c1-e346-4c89-921d-43f5827d2993" containerName="keystone-bootstrap" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.481959 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d056c1-e346-4c89-921d-43f5827d2993" containerName="keystone-bootstrap" Dec 09 17:27:22 crc kubenswrapper[4954]: E1209 17:27:22.481987 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d055d0c-efc9-4476-abed-e27c50e6372a" containerName="init" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.481994 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d055d0c-efc9-4476-abed-e27c50e6372a" containerName="init" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.482198 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d055d0c-efc9-4476-abed-e27c50e6372a" containerName="init" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.482213 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d056c1-e346-4c89-921d-43f5827d2993" containerName="keystone-bootstrap" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.482223 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc313343-7599-4dad-bda6-ecdb04007e4f" containerName="init" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.483157 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.486783 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.486978 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5xlkl" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.486975 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.487106 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.487317 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.496823 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-cb6lc"] Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.527553 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-fernet-keys\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.527708 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngnk5\" (UniqueName: \"kubernetes.io/projected/d0d58b11-2dce-48e1-9445-1c30c431b731-kube-api-access-ngnk5\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.527764 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-combined-ca-bundle\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.527820 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-scripts\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.527868 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-config-data\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.528028 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-credential-keys\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.632953 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-fernet-keys\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.633028 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngnk5\" (UniqueName: \"kubernetes.io/projected/d0d58b11-2dce-48e1-9445-1c30c431b731-kube-api-access-ngnk5\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.633065 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-combined-ca-bundle\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.633104 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-scripts\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.633135 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-config-data\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.633214 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-credential-keys\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.656743 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-config-data\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.658994 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-credential-keys\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.660392 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-combined-ca-bundle\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.661712 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-fernet-keys\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.664431 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-scripts\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.669161 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngnk5\" (UniqueName: \"kubernetes.io/projected/d0d58b11-2dce-48e1-9445-1c30c431b731-kube-api-access-ngnk5\") pod \"keystone-bootstrap-cb6lc\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:22 crc kubenswrapper[4954]: I1209 17:27:22.811273 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:24 crc kubenswrapper[4954]: I1209 17:27:24.135451 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d056c1-e346-4c89-921d-43f5827d2993" path="/var/lib/kubelet/pods/53d056c1-e346-4c89-921d-43f5827d2993/volumes" Dec 09 17:27:24 crc kubenswrapper[4954]: I1209 17:27:24.209136 4954 generic.go:334] "Generic (PLEG): container finished" podID="4296b9c8-795b-4786-80a8-1b8c8d723451" containerID="4104c8ec76bad3c56a6c514d99d167e0bb6322b543aa3dbff81c00a9d841ec15" exitCode=0 Dec 09 17:27:24 crc kubenswrapper[4954]: I1209 17:27:24.209206 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cqb6p" event={"ID":"4296b9c8-795b-4786-80a8-1b8c8d723451","Type":"ContainerDied","Data":"4104c8ec76bad3c56a6c514d99d167e0bb6322b543aa3dbff81c00a9d841ec15"} Dec 09 17:27:28 crc kubenswrapper[4954]: I1209 17:27:28.051809 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-q85b8" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: i/o timeout" Dec 09 17:27:28 crc kubenswrapper[4954]: I1209 17:27:28.052908 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:27:29 crc kubenswrapper[4954]: E1209 17:27:29.761160 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 09 17:27:29 crc kubenswrapper[4954]: E1209 17:27:29.761885 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5wp75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-p5ndh_openstack(25667d18-6643-4344-9ac3-e3a4f1af6957): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:27:29 crc kubenswrapper[4954]: E1209 17:27:29.763492 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-p5ndh" podUID="25667d18-6643-4344-9ac3-e3a4f1af6957" Dec 09 17:27:29 crc kubenswrapper[4954]: I1209 17:27:29.860195 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:27:29 crc kubenswrapper[4954]: I1209 17:27:29.867711 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:27:29 crc kubenswrapper[4954]: I1209 17:27:29.997712 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-sb\") pod \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " Dec 09 17:27:29 crc kubenswrapper[4954]: I1209 17:27:29.998085 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvvct\" (UniqueName: \"kubernetes.io/projected/7801abe0-4c3f-4251-8b29-e8ff3e11b580-kube-api-access-bvvct\") pod \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " Dec 09 17:27:29 crc kubenswrapper[4954]: I1209 17:27:29.998184 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-dns-svc\") pod \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " Dec 09 17:27:29 crc kubenswrapper[4954]: I1209 17:27:29.998225 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-config\") pod \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " Dec 09 17:27:29 crc kubenswrapper[4954]: I1209 17:27:29.998289 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-nb\") pod \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\" (UID: \"7801abe0-4c3f-4251-8b29-e8ff3e11b580\") " Dec 09 17:27:29 crc kubenswrapper[4954]: I1209 17:27:29.998325 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-combined-ca-bundle\") pod \"4296b9c8-795b-4786-80a8-1b8c8d723451\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " Dec 09 17:27:29 crc kubenswrapper[4954]: I1209 17:27:29.998416 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-config\") pod \"4296b9c8-795b-4786-80a8-1b8c8d723451\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " Dec 09 17:27:29 crc kubenswrapper[4954]: I1209 17:27:29.998516 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcnzf\" (UniqueName: \"kubernetes.io/projected/4296b9c8-795b-4786-80a8-1b8c8d723451-kube-api-access-xcnzf\") pod \"4296b9c8-795b-4786-80a8-1b8c8d723451\" (UID: \"4296b9c8-795b-4786-80a8-1b8c8d723451\") " Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.004998 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4296b9c8-795b-4786-80a8-1b8c8d723451-kube-api-access-xcnzf" (OuterVolumeSpecName: "kube-api-access-xcnzf") pod "4296b9c8-795b-4786-80a8-1b8c8d723451" (UID: "4296b9c8-795b-4786-80a8-1b8c8d723451"). InnerVolumeSpecName "kube-api-access-xcnzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.005129 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7801abe0-4c3f-4251-8b29-e8ff3e11b580-kube-api-access-bvvct" (OuterVolumeSpecName: "kube-api-access-bvvct") pod "7801abe0-4c3f-4251-8b29-e8ff3e11b580" (UID: "7801abe0-4c3f-4251-8b29-e8ff3e11b580"). InnerVolumeSpecName "kube-api-access-bvvct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.032982 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-config" (OuterVolumeSpecName: "config") pod "4296b9c8-795b-4786-80a8-1b8c8d723451" (UID: "4296b9c8-795b-4786-80a8-1b8c8d723451"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.051746 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4296b9c8-795b-4786-80a8-1b8c8d723451" (UID: "4296b9c8-795b-4786-80a8-1b8c8d723451"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.057242 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7801abe0-4c3f-4251-8b29-e8ff3e11b580" (UID: "7801abe0-4c3f-4251-8b29-e8ff3e11b580"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.058086 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7801abe0-4c3f-4251-8b29-e8ff3e11b580" (UID: "7801abe0-4c3f-4251-8b29-e8ff3e11b580"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.069646 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-config" (OuterVolumeSpecName: "config") pod "7801abe0-4c3f-4251-8b29-e8ff3e11b580" (UID: "7801abe0-4c3f-4251-8b29-e8ff3e11b580"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.072429 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7801abe0-4c3f-4251-8b29-e8ff3e11b580" (UID: "7801abe0-4c3f-4251-8b29-e8ff3e11b580"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.101764 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.101810 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcnzf\" (UniqueName: \"kubernetes.io/projected/4296b9c8-795b-4786-80a8-1b8c8d723451-kube-api-access-xcnzf\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.101825 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.101835 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvvct\" (UniqueName: \"kubernetes.io/projected/7801abe0-4c3f-4251-8b29-e8ff3e11b580-kube-api-access-bvvct\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.101844 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.101855 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.101865 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7801abe0-4c3f-4251-8b29-e8ff3e11b580-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.101874 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4296b9c8-795b-4786-80a8-1b8c8d723451-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.121001 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:27:30 crc kubenswrapper[4954]: E1209 17:27:30.121325 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.271248 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-q85b8" event={"ID":"7801abe0-4c3f-4251-8b29-e8ff3e11b580","Type":"ContainerDied","Data":"7796d7e5f5ca2ff847c3a82e3ef6f44f70d21d3762d612de78ef6c24ecefb29a"} Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.271286 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-q85b8" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.271318 4954 scope.go:117] "RemoveContainer" containerID="36b7c152baf598616164b052b73516434a385be71cac9598195a5390df8eb505" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.278464 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cqb6p" event={"ID":"4296b9c8-795b-4786-80a8-1b8c8d723451","Type":"ContainerDied","Data":"771f2ca1e031f1cb318bea0fee93a7aa26646a4c4e2c814b8b80635ddc42182f"} Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.278511 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="771f2ca1e031f1cb318bea0fee93a7aa26646a4c4e2c814b8b80635ddc42182f" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.278480 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cqb6p" Dec 09 17:27:30 crc kubenswrapper[4954]: E1209 17:27:30.279886 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-p5ndh" podUID="25667d18-6643-4344-9ac3-e3a4f1af6957" Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.304397 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-q85b8"] Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.317568 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-q85b8"] Dec 09 17:27:30 crc kubenswrapper[4954]: I1209 17:27:30.903844 4954 scope.go:117] "RemoveContainer" containerID="bb0cef67501194586516c8f80b032e6a3bbfae73640e71f342fd1ac48279bdaa" Dec 09 17:27:30 crc kubenswrapper[4954]: E1209 17:27:30.924954 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 09 17:27:30 crc kubenswrapper[4954]: E1209 17:27:30.925387 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l4kr6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-r98b8_openstack(9653f9d1-0297-458e-bfb6-b6801ef96c3c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 17:27:30 crc kubenswrapper[4954]: E1209 17:27:30.926665 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-r98b8" podUID="9653f9d1-0297-458e-bfb6-b6801ef96c3c" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.266570 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-c46lc"] Dec 09 17:27:31 crc kubenswrapper[4954]: E1209 17:27:31.270081 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4296b9c8-795b-4786-80a8-1b8c8d723451" containerName="neutron-db-sync" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.270166 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="4296b9c8-795b-4786-80a8-1b8c8d723451" containerName="neutron-db-sync" Dec 09 17:27:31 crc kubenswrapper[4954]: E1209 17:27:31.270234 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerName="dnsmasq-dns" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.270244 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerName="dnsmasq-dns" Dec 09 17:27:31 crc kubenswrapper[4954]: E1209 17:27:31.270262 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerName="init" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.270279 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerName="init" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.270756 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerName="dnsmasq-dns" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.270807 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="4296b9c8-795b-4786-80a8-1b8c8d723451" containerName="neutron-db-sync" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.272825 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: E1209 17:27:31.318761 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-r98b8" podUID="9653f9d1-0297-458e-bfb6-b6801ef96c3c" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.370558 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-c46lc"] Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.375516 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-svc\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.375820 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9487h\" (UniqueName: \"kubernetes.io/projected/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-kube-api-access-9487h\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.376017 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-config\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.376859 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.377198 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.377313 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.433675 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-76bd5dc95b-mwznb"] Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.437553 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.440312 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.440500 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.440675 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.441071 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fd8bh" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.455158 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76bd5dc95b-mwznb"] Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.466510 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-cb6lc"] Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.478715 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.479649 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twk69\" (UniqueName: \"kubernetes.io/projected/284a949a-68ec-4470-a956-6e0a396675e0-kube-api-access-twk69\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.479756 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-ovndb-tls-certs\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.479857 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.479988 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.480145 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-svc\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.480285 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9487h\" (UniqueName: \"kubernetes.io/projected/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-kube-api-access-9487h\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.480382 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-config\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.480661 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-config\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.480848 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.480882 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-combined-ca-bundle\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.480912 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-httpd-config\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.480987 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-svc\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.481408 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.482017 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-config\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.482118 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.510994 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9487h\" (UniqueName: \"kubernetes.io/projected/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-kube-api-access-9487h\") pod \"dnsmasq-dns-55f844cf75-c46lc\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.582565 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-config\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.582743 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-combined-ca-bundle\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.582764 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-httpd-config\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.582832 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twk69\" (UniqueName: \"kubernetes.io/projected/284a949a-68ec-4470-a956-6e0a396675e0-kube-api-access-twk69\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.582850 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-ovndb-tls-certs\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.588385 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-config\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.588398 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-combined-ca-bundle\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.589010 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-ovndb-tls-certs\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.600989 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-httpd-config\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.606811 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twk69\" (UniqueName: \"kubernetes.io/projected/284a949a-68ec-4470-a956-6e0a396675e0-kube-api-access-twk69\") pod \"neutron-76bd5dc95b-mwznb\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.630783 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:31 crc kubenswrapper[4954]: I1209 17:27:31.760951 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:32 crc kubenswrapper[4954]: I1209 17:27:32.155798 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" path="/var/lib/kubelet/pods/7801abe0-4c3f-4251-8b29-e8ff3e11b580/volumes" Dec 09 17:27:32 crc kubenswrapper[4954]: I1209 17:27:32.339744 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cb6lc" event={"ID":"d0d58b11-2dce-48e1-9445-1c30c431b731","Type":"ContainerStarted","Data":"6f15833686a41166b1a2eb511c9b98157fe8ab2528ceef843f2cbd48ae52828b"} Dec 09 17:27:32 crc kubenswrapper[4954]: I1209 17:27:32.454905 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-c46lc"] Dec 09 17:27:32 crc kubenswrapper[4954]: I1209 17:27:32.663825 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76bd5dc95b-mwznb"] Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.055566 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-q85b8" podUID="7801abe0-4c3f-4251-8b29-e8ff3e11b580" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: i/o timeout" Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.355369 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76bd5dc95b-mwznb" event={"ID":"284a949a-68ec-4470-a956-6e0a396675e0","Type":"ContainerStarted","Data":"58892b864944f9d101cc233d21c8b6e6f2e20e90c186a670c72216c198b1c92e"} Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.355763 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76bd5dc95b-mwznb" event={"ID":"284a949a-68ec-4470-a956-6e0a396675e0","Type":"ContainerStarted","Data":"034e1ffc00381cea4cb662dba60369455c88d5efd04e71909d2f3c1fd0806f78"} Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.358656 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"565df52a-04b4-463b-a1b1-59eee6d94ea5","Type":"ContainerStarted","Data":"69758858f95933f87c890e8522b414326890401c5401220544f9a87b466ba635"} Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.360421 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cb6lc" event={"ID":"d0d58b11-2dce-48e1-9445-1c30c431b731","Type":"ContainerStarted","Data":"8eaf123710b9afd215b706b5e73e57f168af7d51f06ec2b030a11171460b9ffa"} Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.363188 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fsxd4" event={"ID":"5fd76337-c1f2-43ff-b021-8837dd046dc2","Type":"ContainerStarted","Data":"bc9fa86c3f1653c3b63b8f3213022e39348d9149dae183557f2a8bed522b9738"} Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.366958 4954 generic.go:334] "Generic (PLEG): container finished" podID="2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" containerID="26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95" exitCode=0 Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.367055 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" event={"ID":"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb","Type":"ContainerDied","Data":"26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95"} Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.367111 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" event={"ID":"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb","Type":"ContainerStarted","Data":"4aa913ef4bb05c88a19e633a93556dee8a086c4b6d63dc902898d54fbef5ee88"} Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.370248 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c686v" event={"ID":"3b087316-6f76-4aaa-bd43-d5b01051d4b2","Type":"ContainerStarted","Data":"3ed51eac127f0f876dc383fb126166c8f78932579d8696ec733514adecd2268a"} Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.396631 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-cb6lc" podStartSLOduration=11.39660838 podStartE2EDuration="11.39660838s" podCreationTimestamp="2025-12-09 17:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:33.386574477 +0000 UTC m=+1849.774748307" watchObservedRunningTime="2025-12-09 17:27:33.39660838 +0000 UTC m=+1849.784782200" Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.425684 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-fsxd4" podStartSLOduration=6.299326962 podStartE2EDuration="34.425637609s" podCreationTimestamp="2025-12-09 17:26:59 +0000 UTC" firstStartedPulling="2025-12-09 17:27:02.766455789 +0000 UTC m=+1819.154629609" lastFinishedPulling="2025-12-09 17:27:30.892766436 +0000 UTC m=+1847.280940256" observedRunningTime="2025-12-09 17:27:33.411491577 +0000 UTC m=+1849.799665397" watchObservedRunningTime="2025-12-09 17:27:33.425637609 +0000 UTC m=+1849.813811449" Dec 09 17:27:33 crc kubenswrapper[4954]: I1209 17:27:33.440741 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-c686v" podStartSLOduration=5.156196445 podStartE2EDuration="34.440719241s" podCreationTimestamp="2025-12-09 17:26:59 +0000 UTC" firstStartedPulling="2025-12-09 17:27:02.729920655 +0000 UTC m=+1819.118094475" lastFinishedPulling="2025-12-09 17:27:32.014443451 +0000 UTC m=+1848.402617271" observedRunningTime="2025-12-09 17:27:33.435503268 +0000 UTC m=+1849.823677098" watchObservedRunningTime="2025-12-09 17:27:33.440719241 +0000 UTC m=+1849.828893061" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.023426 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9855f74b5-lqsx6"] Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.041965 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.056717 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.061850 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.079692 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9855f74b5-lqsx6"] Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.090160 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-internal-tls-certs\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.090497 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-ovndb-tls-certs\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.090983 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-combined-ca-bundle\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.091220 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4rmr\" (UniqueName: \"kubernetes.io/projected/e52dc021-9970-4cec-8e66-beb3b49a1dd5-kube-api-access-p4rmr\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.091343 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-public-tls-certs\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.091487 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-config\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.091659 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-httpd-config\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.198581 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-internal-tls-certs\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.198657 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-ovndb-tls-certs\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.198767 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-combined-ca-bundle\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.198811 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4rmr\" (UniqueName: \"kubernetes.io/projected/e52dc021-9970-4cec-8e66-beb3b49a1dd5-kube-api-access-p4rmr\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.198859 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-public-tls-certs\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.198903 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-config\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.198938 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-httpd-config\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.204576 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-internal-tls-certs\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.205698 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-config\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.207267 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-public-tls-certs\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.208856 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-httpd-config\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.209198 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-combined-ca-bundle\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.211656 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e52dc021-9970-4cec-8e66-beb3b49a1dd5-ovndb-tls-certs\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.221975 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4rmr\" (UniqueName: \"kubernetes.io/projected/e52dc021-9970-4cec-8e66-beb3b49a1dd5-kube-api-access-p4rmr\") pod \"neutron-9855f74b5-lqsx6\" (UID: \"e52dc021-9970-4cec-8e66-beb3b49a1dd5\") " pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:34 crc kubenswrapper[4954]: I1209 17:27:34.413116 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:35 crc kubenswrapper[4954]: I1209 17:27:35.052896 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9855f74b5-lqsx6"] Dec 09 17:27:35 crc kubenswrapper[4954]: I1209 17:27:35.394122 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9855f74b5-lqsx6" event={"ID":"e52dc021-9970-4cec-8e66-beb3b49a1dd5","Type":"ContainerStarted","Data":"cf72fa0f312561041a101b07f2ebc4fe1f12e094953384bc32764cfe436efb10"} Dec 09 17:27:36 crc kubenswrapper[4954]: I1209 17:27:36.406135 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76bd5dc95b-mwznb" event={"ID":"284a949a-68ec-4470-a956-6e0a396675e0","Type":"ContainerStarted","Data":"67698af236753d0e979ade240e6b3c96329e70ec61deb85d7206df9b666141d6"} Dec 09 17:27:37 crc kubenswrapper[4954]: I1209 17:27:37.423044 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" event={"ID":"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb","Type":"ContainerStarted","Data":"f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af"} Dec 09 17:27:37 crc kubenswrapper[4954]: I1209 17:27:37.423156 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:27:37 crc kubenswrapper[4954]: I1209 17:27:37.462436 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-76bd5dc95b-mwznb" podStartSLOduration=6.462412363 podStartE2EDuration="6.462412363s" podCreationTimestamp="2025-12-09 17:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:37.448776736 +0000 UTC m=+1853.836950556" watchObservedRunningTime="2025-12-09 17:27:37.462412363 +0000 UTC m=+1853.850586183" Dec 09 17:27:38 crc kubenswrapper[4954]: I1209 17:27:38.441492 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9855f74b5-lqsx6" event={"ID":"e52dc021-9970-4cec-8e66-beb3b49a1dd5","Type":"ContainerStarted","Data":"2920e8290ce60acc72bc28f75f48742161c65b45b3ad3f3315652dbcf4f3660c"} Dec 09 17:27:38 crc kubenswrapper[4954]: I1209 17:27:38.442156 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:27:38 crc kubenswrapper[4954]: I1209 17:27:38.442176 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9855f74b5-lqsx6" event={"ID":"e52dc021-9970-4cec-8e66-beb3b49a1dd5","Type":"ContainerStarted","Data":"201ba9de110a588547b4d7f6826e4fc3c2a5f47752a13bff47a3533fba5222e0"} Dec 09 17:27:38 crc kubenswrapper[4954]: I1209 17:27:38.469559 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-9855f74b5-lqsx6" podStartSLOduration=5.469539971 podStartE2EDuration="5.469539971s" podCreationTimestamp="2025-12-09 17:27:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:38.465161484 +0000 UTC m=+1854.853335304" watchObservedRunningTime="2025-12-09 17:27:38.469539971 +0000 UTC m=+1854.857713791" Dec 09 17:27:38 crc kubenswrapper[4954]: I1209 17:27:38.494286 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" podStartSLOduration=7.494262425 podStartE2EDuration="7.494262425s" podCreationTimestamp="2025-12-09 17:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:38.490387984 +0000 UTC m=+1854.878561804" watchObservedRunningTime="2025-12-09 17:27:38.494262425 +0000 UTC m=+1854.882436245" Dec 09 17:27:40 crc kubenswrapper[4954]: I1209 17:27:40.466842 4954 generic.go:334] "Generic (PLEG): container finished" podID="d0d58b11-2dce-48e1-9445-1c30c431b731" containerID="8eaf123710b9afd215b706b5e73e57f168af7d51f06ec2b030a11171460b9ffa" exitCode=0 Dec 09 17:27:40 crc kubenswrapper[4954]: I1209 17:27:40.466901 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cb6lc" event={"ID":"d0d58b11-2dce-48e1-9445-1c30c431b731","Type":"ContainerDied","Data":"8eaf123710b9afd215b706b5e73e57f168af7d51f06ec2b030a11171460b9ffa"} Dec 09 17:27:41 crc kubenswrapper[4954]: I1209 17:27:41.482504 4954 generic.go:334] "Generic (PLEG): container finished" podID="5fd76337-c1f2-43ff-b021-8837dd046dc2" containerID="bc9fa86c3f1653c3b63b8f3213022e39348d9149dae183557f2a8bed522b9738" exitCode=0 Dec 09 17:27:41 crc kubenswrapper[4954]: I1209 17:27:41.482582 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fsxd4" event={"ID":"5fd76337-c1f2-43ff-b021-8837dd046dc2","Type":"ContainerDied","Data":"bc9fa86c3f1653c3b63b8f3213022e39348d9149dae183557f2a8bed522b9738"} Dec 09 17:27:41 crc kubenswrapper[4954]: I1209 17:27:41.632422 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.498146 4954 generic.go:334] "Generic (PLEG): container finished" podID="3b087316-6f76-4aaa-bd43-d5b01051d4b2" containerID="3ed51eac127f0f876dc383fb126166c8f78932579d8696ec733514adecd2268a" exitCode=0 Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.498251 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c686v" event={"ID":"3b087316-6f76-4aaa-bd43-d5b01051d4b2","Type":"ContainerDied","Data":"3ed51eac127f0f876dc383fb126166c8f78932579d8696ec733514adecd2268a"} Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.503260 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cb6lc" event={"ID":"d0d58b11-2dce-48e1-9445-1c30c431b731","Type":"ContainerDied","Data":"6f15833686a41166b1a2eb511c9b98157fe8ab2528ceef843f2cbd48ae52828b"} Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.503322 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f15833686a41166b1a2eb511c9b98157fe8ab2528ceef843f2cbd48ae52828b" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.736014 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.810259 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngnk5\" (UniqueName: \"kubernetes.io/projected/d0d58b11-2dce-48e1-9445-1c30c431b731-kube-api-access-ngnk5\") pod \"d0d58b11-2dce-48e1-9445-1c30c431b731\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.810392 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-fernet-keys\") pod \"d0d58b11-2dce-48e1-9445-1c30c431b731\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.810569 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-combined-ca-bundle\") pod \"d0d58b11-2dce-48e1-9445-1c30c431b731\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.810656 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-scripts\") pod \"d0d58b11-2dce-48e1-9445-1c30c431b731\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.810740 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-config-data\") pod \"d0d58b11-2dce-48e1-9445-1c30c431b731\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.810864 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-credential-keys\") pod \"d0d58b11-2dce-48e1-9445-1c30c431b731\" (UID: \"d0d58b11-2dce-48e1-9445-1c30c431b731\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.815176 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-scripts" (OuterVolumeSpecName: "scripts") pod "d0d58b11-2dce-48e1-9445-1c30c431b731" (UID: "d0d58b11-2dce-48e1-9445-1c30c431b731"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.815666 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d58b11-2dce-48e1-9445-1c30c431b731-kube-api-access-ngnk5" (OuterVolumeSpecName: "kube-api-access-ngnk5") pod "d0d58b11-2dce-48e1-9445-1c30c431b731" (UID: "d0d58b11-2dce-48e1-9445-1c30c431b731"). InnerVolumeSpecName "kube-api-access-ngnk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.816732 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d0d58b11-2dce-48e1-9445-1c30c431b731" (UID: "d0d58b11-2dce-48e1-9445-1c30c431b731"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.822989 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d0d58b11-2dce-48e1-9445-1c30c431b731" (UID: "d0d58b11-2dce-48e1-9445-1c30c431b731"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.844254 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-config-data" (OuterVolumeSpecName: "config-data") pod "d0d58b11-2dce-48e1-9445-1c30c431b731" (UID: "d0d58b11-2dce-48e1-9445-1c30c431b731"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.854250 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0d58b11-2dce-48e1-9445-1c30c431b731" (UID: "d0d58b11-2dce-48e1-9445-1c30c431b731"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.876896 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.912456 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-scripts\") pod \"5fd76337-c1f2-43ff-b021-8837dd046dc2\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.913028 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp2f2\" (UniqueName: \"kubernetes.io/projected/5fd76337-c1f2-43ff-b021-8837dd046dc2-kube-api-access-wp2f2\") pod \"5fd76337-c1f2-43ff-b021-8837dd046dc2\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.913142 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-combined-ca-bundle\") pod \"5fd76337-c1f2-43ff-b021-8837dd046dc2\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.913370 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fd76337-c1f2-43ff-b021-8837dd046dc2-logs\") pod \"5fd76337-c1f2-43ff-b021-8837dd046dc2\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.913419 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-config-data\") pod \"5fd76337-c1f2-43ff-b021-8837dd046dc2\" (UID: \"5fd76337-c1f2-43ff-b021-8837dd046dc2\") " Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.913755 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fd76337-c1f2-43ff-b021-8837dd046dc2-logs" (OuterVolumeSpecName: "logs") pod "5fd76337-c1f2-43ff-b021-8837dd046dc2" (UID: "5fd76337-c1f2-43ff-b021-8837dd046dc2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.914405 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.914439 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.914455 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.914468 4954 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.914481 4954 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fd76337-c1f2-43ff-b021-8837dd046dc2-logs\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.914493 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngnk5\" (UniqueName: \"kubernetes.io/projected/d0d58b11-2dce-48e1-9445-1c30c431b731-kube-api-access-ngnk5\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.914506 4954 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0d58b11-2dce-48e1-9445-1c30c431b731-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.915755 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-scripts" (OuterVolumeSpecName: "scripts") pod "5fd76337-c1f2-43ff-b021-8837dd046dc2" (UID: "5fd76337-c1f2-43ff-b021-8837dd046dc2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.925310 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd76337-c1f2-43ff-b021-8837dd046dc2-kube-api-access-wp2f2" (OuterVolumeSpecName: "kube-api-access-wp2f2") pod "5fd76337-c1f2-43ff-b021-8837dd046dc2" (UID: "5fd76337-c1f2-43ff-b021-8837dd046dc2"). InnerVolumeSpecName "kube-api-access-wp2f2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.947880 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-config-data" (OuterVolumeSpecName: "config-data") pod "5fd76337-c1f2-43ff-b021-8837dd046dc2" (UID: "5fd76337-c1f2-43ff-b021-8837dd046dc2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:42 crc kubenswrapper[4954]: I1209 17:27:42.956116 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fd76337-c1f2-43ff-b021-8837dd046dc2" (UID: "5fd76337-c1f2-43ff-b021-8837dd046dc2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.016484 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.016519 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.016529 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp2f2\" (UniqueName: \"kubernetes.io/projected/5fd76337-c1f2-43ff-b021-8837dd046dc2-kube-api-access-wp2f2\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.016542 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd76337-c1f2-43ff-b021-8837dd046dc2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.517081 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fsxd4" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.518274 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fsxd4" event={"ID":"5fd76337-c1f2-43ff-b021-8837dd046dc2","Type":"ContainerDied","Data":"d2eeaa2edb192357d38d5aef2930c8b5dc195ec06ca87a6975ce9d748377c000"} Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.518363 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2eeaa2edb192357d38d5aef2930c8b5dc195ec06ca87a6975ce9d748377c000" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.543923 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r98b8" event={"ID":"9653f9d1-0297-458e-bfb6-b6801ef96c3c","Type":"ContainerStarted","Data":"9ce2957549eff70070e4dc18369586f0756e0c26b3c09557904e5f37ab26085c"} Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.552225 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"565df52a-04b4-463b-a1b1-59eee6d94ea5","Type":"ContainerStarted","Data":"618ff916e034b8d1cc36791523390bc905f827eb4053d62b5486e49cb605cfc4"} Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.552308 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cb6lc" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.623216 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-r98b8" podStartSLOduration=4.617359377 podStartE2EDuration="44.623179228s" podCreationTimestamp="2025-12-09 17:26:59 +0000 UTC" firstStartedPulling="2025-12-09 17:27:02.77573555 +0000 UTC m=+1819.163909370" lastFinishedPulling="2025-12-09 17:27:42.781555401 +0000 UTC m=+1859.169729221" observedRunningTime="2025-12-09 17:27:43.568981282 +0000 UTC m=+1859.957155102" watchObservedRunningTime="2025-12-09 17:27:43.623179228 +0000 UTC m=+1860.011353048" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.659986 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-57bb47ccd8-dn5h5"] Dec 09 17:27:43 crc kubenswrapper[4954]: E1209 17:27:43.660634 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d58b11-2dce-48e1-9445-1c30c431b731" containerName="keystone-bootstrap" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.660656 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d58b11-2dce-48e1-9445-1c30c431b731" containerName="keystone-bootstrap" Dec 09 17:27:43 crc kubenswrapper[4954]: E1209 17:27:43.660682 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd76337-c1f2-43ff-b021-8837dd046dc2" containerName="placement-db-sync" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.660690 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd76337-c1f2-43ff-b021-8837dd046dc2" containerName="placement-db-sync" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.660933 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd76337-c1f2-43ff-b021-8837dd046dc2" containerName="placement-db-sync" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.660972 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d58b11-2dce-48e1-9445-1c30c431b731" containerName="keystone-bootstrap" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.662986 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.666541 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.667200 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-4ztm9" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.667801 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.669528 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.670041 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.682628 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-57bb47ccd8-dn5h5"] Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.736171 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-config-data\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.736277 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rw26\" (UniqueName: \"kubernetes.io/projected/ce0dd183-082d-42f3-8e74-050349e431fa-kube-api-access-5rw26\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.736326 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-combined-ca-bundle\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.736362 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-public-tls-certs\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.736432 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-internal-tls-certs\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.736467 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce0dd183-082d-42f3-8e74-050349e431fa-logs\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.736503 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-scripts\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.838022 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-config-data\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.838514 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rw26\" (UniqueName: \"kubernetes.io/projected/ce0dd183-082d-42f3-8e74-050349e431fa-kube-api-access-5rw26\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.838643 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-combined-ca-bundle\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.838723 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-public-tls-certs\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.838824 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-internal-tls-certs\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.838908 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce0dd183-082d-42f3-8e74-050349e431fa-logs\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.840454 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce0dd183-082d-42f3-8e74-050349e431fa-logs\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.848506 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-config-data\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.848823 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-public-tls-certs\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.848834 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-combined-ca-bundle\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.850914 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-scripts\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.861886 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-scripts\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.868297 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce0dd183-082d-42f3-8e74-050349e431fa-internal-tls-certs\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.869555 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rw26\" (UniqueName: \"kubernetes.io/projected/ce0dd183-082d-42f3-8e74-050349e431fa-kube-api-access-5rw26\") pod \"placement-57bb47ccd8-dn5h5\" (UID: \"ce0dd183-082d-42f3-8e74-050349e431fa\") " pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.941447 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-64ffb74c6f-59w54"] Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.943054 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.954909 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.955199 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.955355 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.955524 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5xlkl" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.955719 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.955796 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.965748 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64ffb74c6f-59w54"] Dec 09 17:27:43 crc kubenswrapper[4954]: I1209 17:27:43.989309 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.059001 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-credential-keys\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.059990 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-scripts\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.060202 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-config-data\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.060309 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-combined-ca-bundle\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.060456 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tdvz\" (UniqueName: \"kubernetes.io/projected/4e460517-3e4e-4179-b4f0-1b1db948715b-kube-api-access-9tdvz\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.060770 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-fernet-keys\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.060967 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-public-tls-certs\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.061069 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-internal-tls-certs\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.133280 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:27:44 crc kubenswrapper[4954]: E1209 17:27:44.133638 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.163854 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-config-data\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.164227 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-combined-ca-bundle\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.164330 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tdvz\" (UniqueName: \"kubernetes.io/projected/4e460517-3e4e-4179-b4f0-1b1db948715b-kube-api-access-9tdvz\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.164464 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-fernet-keys\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.164748 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-public-tls-certs\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.164868 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-internal-tls-certs\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.165004 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-credential-keys\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.165101 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-scripts\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.167790 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.170581 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.170816 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.170839 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.171136 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.179383 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.180265 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-config-data\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.185694 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-scripts\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.185863 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-credential-keys\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.185996 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-combined-ca-bundle\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.190131 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-internal-tls-certs\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.190792 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tdvz\" (UniqueName: \"kubernetes.io/projected/4e460517-3e4e-4179-b4f0-1b1db948715b-kube-api-access-9tdvz\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.198280 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-public-tls-certs\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.209070 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e460517-3e4e-4179-b4f0-1b1db948715b-fernet-keys\") pod \"keystone-64ffb74c6f-59w54\" (UID: \"4e460517-3e4e-4179-b4f0-1b1db948715b\") " pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.267924 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-combined-ca-bundle\") pod \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.268079 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq92f\" (UniqueName: \"kubernetes.io/projected/3b087316-6f76-4aaa-bd43-d5b01051d4b2-kube-api-access-tq92f\") pod \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.268268 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-db-sync-config-data\") pod \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\" (UID: \"3b087316-6f76-4aaa-bd43-d5b01051d4b2\") " Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.272740 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3b087316-6f76-4aaa-bd43-d5b01051d4b2" (UID: "3b087316-6f76-4aaa-bd43-d5b01051d4b2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.273309 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b087316-6f76-4aaa-bd43-d5b01051d4b2-kube-api-access-tq92f" (OuterVolumeSpecName: "kube-api-access-tq92f") pod "3b087316-6f76-4aaa-bd43-d5b01051d4b2" (UID: "3b087316-6f76-4aaa-bd43-d5b01051d4b2"). InnerVolumeSpecName "kube-api-access-tq92f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.298141 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b087316-6f76-4aaa-bd43-d5b01051d4b2" (UID: "3b087316-6f76-4aaa-bd43-d5b01051d4b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.312316 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5xlkl" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.320141 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.372376 4954 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.372417 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b087316-6f76-4aaa-bd43-d5b01051d4b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.372427 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq92f\" (UniqueName: \"kubernetes.io/projected/3b087316-6f76-4aaa-bd43-d5b01051d4b2-kube-api-access-tq92f\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.595187 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-57bb47ccd8-dn5h5"] Dec 09 17:27:44 crc kubenswrapper[4954]: W1209 17:27:44.611097 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0dd183_082d_42f3_8e74_050349e431fa.slice/crio-5aa853454f1c5ddeb300a36abd8a3fe2bda93e17adb411fb8bcbb9ddc5ddb848 WatchSource:0}: Error finding container 5aa853454f1c5ddeb300a36abd8a3fe2bda93e17adb411fb8bcbb9ddc5ddb848: Status 404 returned error can't find the container with id 5aa853454f1c5ddeb300a36abd8a3fe2bda93e17adb411fb8bcbb9ddc5ddb848 Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.644355 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-c686v" event={"ID":"3b087316-6f76-4aaa-bd43-d5b01051d4b2","Type":"ContainerDied","Data":"fb492d5eafcaef5a0ec12ac9ebc9ccf2bdec9a40db78134e0c7c7e54bf514204"} Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.644400 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb492d5eafcaef5a0ec12ac9ebc9ccf2bdec9a40db78134e0c7c7e54bf514204" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.644469 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-c686v" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.843313 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-76bccc865f-nls4t"] Dec 09 17:27:44 crc kubenswrapper[4954]: E1209 17:27:44.847868 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b087316-6f76-4aaa-bd43-d5b01051d4b2" containerName="barbican-db-sync" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.847917 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b087316-6f76-4aaa-bd43-d5b01051d4b2" containerName="barbican-db-sync" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.848278 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b087316-6f76-4aaa-bd43-d5b01051d4b2" containerName="barbican-db-sync" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.849794 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.853944 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.866962 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.868210 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-n4cmp" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.887281 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7cbfd68ff6-49k8v"] Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.889496 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.892343 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.898834 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704e4510-1e1b-492f-9de5-e5bbec8e6731-combined-ca-bundle\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.898920 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgwg7\" (UniqueName: \"kubernetes.io/projected/704e4510-1e1b-492f-9de5-e5bbec8e6731-kube-api-access-vgwg7\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.899043 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/704e4510-1e1b-492f-9de5-e5bbec8e6731-config-data-custom\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.899073 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704e4510-1e1b-492f-9de5-e5bbec8e6731-logs\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.899240 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704e4510-1e1b-492f-9de5-e5bbec8e6731-config-data\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.927534 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76bccc865f-nls4t"] Dec 09 17:27:44 crc kubenswrapper[4954]: I1209 17:27:44.991100 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7cbfd68ff6-49k8v"] Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.011657 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704e4510-1e1b-492f-9de5-e5bbec8e6731-config-data\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.011815 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad27e7b2-2f17-4dbc-b611-53f192f502ae-logs\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.012211 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmtsr\" (UniqueName: \"kubernetes.io/projected/ad27e7b2-2f17-4dbc-b611-53f192f502ae-kube-api-access-mmtsr\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.012369 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad27e7b2-2f17-4dbc-b611-53f192f502ae-config-data-custom\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.012430 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704e4510-1e1b-492f-9de5-e5bbec8e6731-combined-ca-bundle\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.012503 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgwg7\" (UniqueName: \"kubernetes.io/projected/704e4510-1e1b-492f-9de5-e5bbec8e6731-kube-api-access-vgwg7\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.012557 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad27e7b2-2f17-4dbc-b611-53f192f502ae-combined-ca-bundle\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.012757 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad27e7b2-2f17-4dbc-b611-53f192f502ae-config-data\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.012859 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/704e4510-1e1b-492f-9de5-e5bbec8e6731-config-data-custom\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.012957 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704e4510-1e1b-492f-9de5-e5bbec8e6731-logs\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.040742 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/704e4510-1e1b-492f-9de5-e5bbec8e6731-logs\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.055748 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/704e4510-1e1b-492f-9de5-e5bbec8e6731-config-data\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.057089 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/704e4510-1e1b-492f-9de5-e5bbec8e6731-combined-ca-bundle\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.058531 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/704e4510-1e1b-492f-9de5-e5bbec8e6731-config-data-custom\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.094895 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgwg7\" (UniqueName: \"kubernetes.io/projected/704e4510-1e1b-492f-9de5-e5bbec8e6731-kube-api-access-vgwg7\") pod \"barbican-worker-76bccc865f-nls4t\" (UID: \"704e4510-1e1b-492f-9de5-e5bbec8e6731\") " pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.118352 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad27e7b2-2f17-4dbc-b611-53f192f502ae-config-data-custom\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.118507 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad27e7b2-2f17-4dbc-b611-53f192f502ae-combined-ca-bundle\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.118738 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad27e7b2-2f17-4dbc-b611-53f192f502ae-config-data\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.119515 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad27e7b2-2f17-4dbc-b611-53f192f502ae-logs\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.120875 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad27e7b2-2f17-4dbc-b611-53f192f502ae-logs\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.121210 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmtsr\" (UniqueName: \"kubernetes.io/projected/ad27e7b2-2f17-4dbc-b611-53f192f502ae-kube-api-access-mmtsr\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.140750 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad27e7b2-2f17-4dbc-b611-53f192f502ae-config-data\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.142726 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmtsr\" (UniqueName: \"kubernetes.io/projected/ad27e7b2-2f17-4dbc-b611-53f192f502ae-kube-api-access-mmtsr\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.158564 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad27e7b2-2f17-4dbc-b611-53f192f502ae-config-data-custom\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.187880 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad27e7b2-2f17-4dbc-b611-53f192f502ae-combined-ca-bundle\") pod \"barbican-keystone-listener-7cbfd68ff6-49k8v\" (UID: \"ad27e7b2-2f17-4dbc-b611-53f192f502ae\") " pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.187987 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64ffb74c6f-59w54"] Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.263888 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76bccc865f-nls4t" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.269922 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-c46lc"] Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.270208 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" podUID="2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" containerName="dnsmasq-dns" containerID="cri-o://f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af" gracePeriod=10 Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.273153 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.306145 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-rsthw"] Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.308303 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.338520 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.351097 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-rsthw"] Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.368800 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-f98894b5d-v8d8l"] Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.371480 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.374865 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.390097 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f98894b5d-v8d8l"] Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.441164 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-config\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.441263 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.441300 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.443175 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.443349 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l8pm\" (UniqueName: \"kubernetes.io/projected/d0fe7343-f624-4a85-a1c2-1c05e4551159-kube-api-access-5l8pm\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.443447 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-svc\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.552865 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data-custom\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.554630 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l8pm\" (UniqueName: \"kubernetes.io/projected/d0fe7343-f624-4a85-a1c2-1c05e4551159-kube-api-access-5l8pm\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.554795 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-svc\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.554917 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.555094 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-config\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.555213 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-logs\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.555313 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.555365 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.555444 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bbxw\" (UniqueName: \"kubernetes.io/projected/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-kube-api-access-5bbxw\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.555521 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.555748 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-combined-ca-bundle\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.556441 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-config\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.556535 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-svc\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.556643 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.557509 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.567393 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.590813 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l8pm\" (UniqueName: \"kubernetes.io/projected/d0fe7343-f624-4a85-a1c2-1c05e4551159-kube-api-access-5l8pm\") pod \"dnsmasq-dns-85ff748b95-rsthw\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.655456 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.657465 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-combined-ca-bundle\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.657557 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data-custom\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.657668 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.657752 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-logs\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.657835 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bbxw\" (UniqueName: \"kubernetes.io/projected/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-kube-api-access-5bbxw\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.658561 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-logs\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.674353 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data-custom\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.677973 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-combined-ca-bundle\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.687371 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bbxw\" (UniqueName: \"kubernetes.io/projected/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-kube-api-access-5bbxw\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.688523 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data\") pod \"barbican-api-f98894b5d-v8d8l\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.690475 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.752835 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57bb47ccd8-dn5h5" event={"ID":"ce0dd183-082d-42f3-8e74-050349e431fa","Type":"ContainerStarted","Data":"5aa853454f1c5ddeb300a36abd8a3fe2bda93e17adb411fb8bcbb9ddc5ddb848"} Dec 09 17:27:45 crc kubenswrapper[4954]: I1209 17:27:45.779214 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64ffb74c6f-59w54" event={"ID":"4e460517-3e4e-4179-b4f0-1b1db948715b","Type":"ContainerStarted","Data":"f5ceeaab42e41e990177383d7e0e73fb6e89a02769a159b735662c6deea79ad2"} Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.093567 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76bccc865f-nls4t"] Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.268209 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7cbfd68ff6-49k8v"] Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.494097 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-rsthw"] Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.821045 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-p5ndh" event={"ID":"25667d18-6643-4344-9ac3-e3a4f1af6957","Type":"ContainerStarted","Data":"75b288f79920a57cf39dd57ed532721dab0f87ebf88519013a2061e8f811e374"} Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.841622 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.847504 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" event={"ID":"ad27e7b2-2f17-4dbc-b611-53f192f502ae","Type":"ContainerStarted","Data":"aeef100c5b8ba1dda94181fb94a79a5f81835c490f24ba01673f6f2f4ea62e45"} Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.851037 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" event={"ID":"d0fe7343-f624-4a85-a1c2-1c05e4551159","Type":"ContainerStarted","Data":"08fc02cd975a0f7b0145fab466c76f6b33b79f9b56f77e48bdb44de07b7e0293"} Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.861408 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-p5ndh" podStartSLOduration=5.088118744 podStartE2EDuration="47.861386452s" podCreationTimestamp="2025-12-09 17:26:59 +0000 UTC" firstStartedPulling="2025-12-09 17:27:02.340839585 +0000 UTC m=+1818.729013405" lastFinishedPulling="2025-12-09 17:27:45.114107293 +0000 UTC m=+1861.502281113" observedRunningTime="2025-12-09 17:27:46.847296001 +0000 UTC m=+1863.235469821" watchObservedRunningTime="2025-12-09 17:27:46.861386452 +0000 UTC m=+1863.249560272" Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.872122 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57bb47ccd8-dn5h5" event={"ID":"ce0dd183-082d-42f3-8e74-050349e431fa","Type":"ContainerStarted","Data":"d1923b6595cc2bda831851c9f043beb16539490f507bcfec82f9eeadc50e580c"} Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.875565 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57bb47ccd8-dn5h5" event={"ID":"ce0dd183-082d-42f3-8e74-050349e431fa","Type":"ContainerStarted","Data":"e85a56dc29d5d8d743105c8ba25fa6d0a2bb0fa2418a4a49a0310472e44a80ac"} Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.875622 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.875643 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.881811 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bccc865f-nls4t" event={"ID":"704e4510-1e1b-492f-9de5-e5bbec8e6731","Type":"ContainerStarted","Data":"65bbd2de371c3f6c9466574d2755f3642adddc92f3d762d3443c3b8a19bede0b"} Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.955279 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-57bb47ccd8-dn5h5" podStartSLOduration=3.955257611 podStartE2EDuration="3.955257611s" podCreationTimestamp="2025-12-09 17:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:46.924107336 +0000 UTC m=+1863.312281176" watchObservedRunningTime="2025-12-09 17:27:46.955257611 +0000 UTC m=+1863.343431431" Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.958716 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64ffb74c6f-59w54" event={"ID":"4e460517-3e4e-4179-b4f0-1b1db948715b","Type":"ContainerStarted","Data":"5ca2dced0c0b72618a88245bcc35fa2e55689b03198e3465506dd57841008253"} Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.960763 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.981924 4954 generic.go:334] "Generic (PLEG): container finished" podID="2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" containerID="f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af" exitCode=0 Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.982927 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" event={"ID":"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb","Type":"ContainerDied","Data":"f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af"} Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.982988 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" event={"ID":"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb","Type":"ContainerDied","Data":"4aa913ef4bb05c88a19e633a93556dee8a086c4b6d63dc902898d54fbef5ee88"} Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.983027 4954 scope.go:117] "RemoveContainer" containerID="f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af" Dec 09 17:27:46 crc kubenswrapper[4954]: I1209 17:27:46.983122 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.017129 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-64ffb74c6f-59w54" podStartSLOduration=4.017090687 podStartE2EDuration="4.017090687s" podCreationTimestamp="2025-12-09 17:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:46.987358305 +0000 UTC m=+1863.375532145" watchObservedRunningTime="2025-12-09 17:27:47.017090687 +0000 UTC m=+1863.405264507" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.038828 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-nb\") pod \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.038904 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-svc\") pod \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.038932 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-sb\") pod \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.039000 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-config\") pod \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.039115 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-swift-storage-0\") pod \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.039176 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9487h\" (UniqueName: \"kubernetes.io/projected/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-kube-api-access-9487h\") pod \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\" (UID: \"2ddac2d7-0d59-4d5f-b08b-41363e4fcadb\") " Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.081359 4954 scope.go:117] "RemoveContainer" containerID="26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.081409 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-kube-api-access-9487h" (OuterVolumeSpecName: "kube-api-access-9487h") pod "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" (UID: "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb"). InnerVolumeSpecName "kube-api-access-9487h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.093875 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f98894b5d-v8d8l"] Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.139732 4954 scope.go:117] "RemoveContainer" containerID="f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af" Dec 09 17:27:47 crc kubenswrapper[4954]: E1209 17:27:47.140284 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af\": container with ID starting with f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af not found: ID does not exist" containerID="f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.140322 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af"} err="failed to get container status \"f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af\": rpc error: code = NotFound desc = could not find container \"f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af\": container with ID starting with f486d4123c244dabbfe1ba1c71941fab552c34da7e9900104f5948f3104ae4af not found: ID does not exist" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.140353 4954 scope.go:117] "RemoveContainer" containerID="26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95" Dec 09 17:27:47 crc kubenswrapper[4954]: E1209 17:27:47.140661 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95\": container with ID starting with 26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95 not found: ID does not exist" containerID="26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.140710 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95"} err="failed to get container status \"26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95\": rpc error: code = NotFound desc = could not find container \"26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95\": container with ID starting with 26d6e70b4dba0f14f579431223b5667fd17eb5821633d22af0f2bd43429b3c95 not found: ID does not exist" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.155639 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9487h\" (UniqueName: \"kubernetes.io/projected/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-kube-api-access-9487h\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.186725 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-config" (OuterVolumeSpecName: "config") pod "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" (UID: "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.195960 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" (UID: "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.207612 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" (UID: "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.209210 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" (UID: "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.216919 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" (UID: "2ddac2d7-0d59-4d5f-b08b-41363e4fcadb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.257758 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.257804 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.257852 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.257868 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.257882 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.382035 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-c46lc"] Dec 09 17:27:47 crc kubenswrapper[4954]: I1209 17:27:47.419085 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-c46lc"] Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.014360 4954 generic.go:334] "Generic (PLEG): container finished" podID="d0fe7343-f624-4a85-a1c2-1c05e4551159" containerID="fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318" exitCode=0 Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.014749 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" event={"ID":"d0fe7343-f624-4a85-a1c2-1c05e4551159","Type":"ContainerDied","Data":"fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318"} Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.022963 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f98894b5d-v8d8l" event={"ID":"9b5721ed-4849-42c5-90c8-5524bcd5d9e1","Type":"ContainerStarted","Data":"1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea"} Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.023039 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f98894b5d-v8d8l" event={"ID":"9b5721ed-4849-42c5-90c8-5524bcd5d9e1","Type":"ContainerStarted","Data":"f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38"} Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.023052 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f98894b5d-v8d8l" event={"ID":"9b5721ed-4849-42c5-90c8-5524bcd5d9e1","Type":"ContainerStarted","Data":"001a939c1e5d78928b10c3058cb99317ab202be26f727ffda3c9089a8da562f6"} Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.080871 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-f98894b5d-v8d8l" podStartSLOduration=3.080846828 podStartE2EDuration="3.080846828s" podCreationTimestamp="2025-12-09 17:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:48.077310047 +0000 UTC m=+1864.465483877" watchObservedRunningTime="2025-12-09 17:27:48.080846828 +0000 UTC m=+1864.469020648" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.138876 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" path="/var/lib/kubelet/pods/2ddac2d7-0d59-4d5f-b08b-41363e4fcadb/volumes" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.614808 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6dfdb467d-r2rqf"] Dec 09 17:27:48 crc kubenswrapper[4954]: E1209 17:27:48.615964 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" containerName="init" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.615994 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" containerName="init" Dec 09 17:27:48 crc kubenswrapper[4954]: E1209 17:27:48.616047 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" containerName="dnsmasq-dns" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.616056 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" containerName="dnsmasq-dns" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.616324 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" containerName="dnsmasq-dns" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.617622 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.621183 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.621467 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.630114 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6dfdb467d-r2rqf"] Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.737017 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22f55\" (UniqueName: \"kubernetes.io/projected/653b2aa3-1e55-42ae-9284-51d651461a15-kube-api-access-22f55\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.737096 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-internal-tls-certs\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.737134 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-config-data-custom\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.737233 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-config-data\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.737265 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-combined-ca-bundle\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.737293 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-public-tls-certs\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.737364 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/653b2aa3-1e55-42ae-9284-51d651461a15-logs\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.839612 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22f55\" (UniqueName: \"kubernetes.io/projected/653b2aa3-1e55-42ae-9284-51d651461a15-kube-api-access-22f55\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.839678 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-internal-tls-certs\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.839711 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-config-data-custom\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.839773 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-config-data\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.839807 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-combined-ca-bundle\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.839838 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-public-tls-certs\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.839903 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/653b2aa3-1e55-42ae-9284-51d651461a15-logs\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.840515 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/653b2aa3-1e55-42ae-9284-51d651461a15-logs\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.846105 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-config-data\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.853156 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-public-tls-certs\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.853901 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-config-data-custom\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.854173 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-combined-ca-bundle\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.855003 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/653b2aa3-1e55-42ae-9284-51d651461a15-internal-tls-certs\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.873347 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22f55\" (UniqueName: \"kubernetes.io/projected/653b2aa3-1e55-42ae-9284-51d651461a15-kube-api-access-22f55\") pod \"barbican-api-6dfdb467d-r2rqf\" (UID: \"653b2aa3-1e55-42ae-9284-51d651461a15\") " pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:48 crc kubenswrapper[4954]: I1209 17:27:48.946785 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:49 crc kubenswrapper[4954]: I1209 17:27:49.051956 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" event={"ID":"d0fe7343-f624-4a85-a1c2-1c05e4551159","Type":"ContainerStarted","Data":"61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d"} Dec 09 17:27:49 crc kubenswrapper[4954]: I1209 17:27:49.052224 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:49 crc kubenswrapper[4954]: I1209 17:27:49.052283 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:49 crc kubenswrapper[4954]: I1209 17:27:49.090842 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" podStartSLOduration=4.090819155 podStartE2EDuration="4.090819155s" podCreationTimestamp="2025-12-09 17:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:49.083088424 +0000 UTC m=+1865.471262254" watchObservedRunningTime="2025-12-09 17:27:49.090819155 +0000 UTC m=+1865.478992975" Dec 09 17:27:50 crc kubenswrapper[4954]: I1209 17:27:50.071664 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:50 crc kubenswrapper[4954]: W1209 17:27:50.321205 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod653b2aa3_1e55_42ae_9284_51d651461a15.slice/crio-13b10923b9c8a13b1c67e6d46e1be9b87933c409226168eb2b1bc98e0e57e22f WatchSource:0}: Error finding container 13b10923b9c8a13b1c67e6d46e1be9b87933c409226168eb2b1bc98e0e57e22f: Status 404 returned error can't find the container with id 13b10923b9c8a13b1c67e6d46e1be9b87933c409226168eb2b1bc98e0e57e22f Dec 09 17:27:50 crc kubenswrapper[4954]: I1209 17:27:50.321371 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6dfdb467d-r2rqf"] Dec 09 17:27:51 crc kubenswrapper[4954]: I1209 17:27:51.089525 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bccc865f-nls4t" event={"ID":"704e4510-1e1b-492f-9de5-e5bbec8e6731","Type":"ContainerStarted","Data":"446463d75b0a8669ff458808991dd251ab6f8e8056ad7e703c5000057f10cf74"} Dec 09 17:27:51 crc kubenswrapper[4954]: I1209 17:27:51.100211 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" event={"ID":"ad27e7b2-2f17-4dbc-b611-53f192f502ae","Type":"ContainerStarted","Data":"8fb6ce2e4bcf071df9580b4e26524062439722f00810d616573d430936477b7d"} Dec 09 17:27:51 crc kubenswrapper[4954]: I1209 17:27:51.104037 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dfdb467d-r2rqf" event={"ID":"653b2aa3-1e55-42ae-9284-51d651461a15","Type":"ContainerStarted","Data":"13b10923b9c8a13b1c67e6d46e1be9b87933c409226168eb2b1bc98e0e57e22f"} Dec 09 17:27:51 crc kubenswrapper[4954]: I1209 17:27:51.131185 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" podStartSLOduration=3.63787363 podStartE2EDuration="7.131164509s" podCreationTimestamp="2025-12-09 17:27:44 +0000 UTC" firstStartedPulling="2025-12-09 17:27:46.272821657 +0000 UTC m=+1862.660995477" lastFinishedPulling="2025-12-09 17:27:49.766112536 +0000 UTC m=+1866.154286356" observedRunningTime="2025-12-09 17:27:51.121700323 +0000 UTC m=+1867.509874163" watchObservedRunningTime="2025-12-09 17:27:51.131164509 +0000 UTC m=+1867.519338329" Dec 09 17:27:51 crc kubenswrapper[4954]: I1209 17:27:51.632804 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-c46lc" podUID="2ddac2d7-0d59-4d5f-b08b-41363e4fcadb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.182:5353: i/o timeout" Dec 09 17:27:52 crc kubenswrapper[4954]: I1209 17:27:52.117190 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dfdb467d-r2rqf" event={"ID":"653b2aa3-1e55-42ae-9284-51d651461a15","Type":"ContainerStarted","Data":"456ebad6895d5f34c5132476e142335bbb0a0c3ce439fd634aba495ab262ddf7"} Dec 09 17:27:52 crc kubenswrapper[4954]: I1209 17:27:52.133150 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bccc865f-nls4t" event={"ID":"704e4510-1e1b-492f-9de5-e5bbec8e6731","Type":"ContainerStarted","Data":"b282eefbfd0c1f0cd93cfc8288f3544534fce02afda352bc575d9291dceaf448"} Dec 09 17:27:52 crc kubenswrapper[4954]: I1209 17:27:52.133214 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cbfd68ff6-49k8v" event={"ID":"ad27e7b2-2f17-4dbc-b611-53f192f502ae","Type":"ContainerStarted","Data":"1eff70aa0911cc31376d3728b92fa2f9e6cd973fd8941933b39fe60715b07b07"} Dec 09 17:27:52 crc kubenswrapper[4954]: I1209 17:27:52.146774 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-76bccc865f-nls4t" podStartSLOduration=4.555095634 podStartE2EDuration="8.146753773s" podCreationTimestamp="2025-12-09 17:27:44 +0000 UTC" firstStartedPulling="2025-12-09 17:27:46.172435834 +0000 UTC m=+1862.560609654" lastFinishedPulling="2025-12-09 17:27:49.764093973 +0000 UTC m=+1866.152267793" observedRunningTime="2025-12-09 17:27:52.139176075 +0000 UTC m=+1868.527349895" watchObservedRunningTime="2025-12-09 17:27:52.146753773 +0000 UTC m=+1868.534927593" Dec 09 17:27:53 crc kubenswrapper[4954]: I1209 17:27:53.146668 4954 generic.go:334] "Generic (PLEG): container finished" podID="25667d18-6643-4344-9ac3-e3a4f1af6957" containerID="75b288f79920a57cf39dd57ed532721dab0f87ebf88519013a2061e8f811e374" exitCode=0 Dec 09 17:27:53 crc kubenswrapper[4954]: I1209 17:27:53.147077 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-p5ndh" event={"ID":"25667d18-6643-4344-9ac3-e3a4f1af6957","Type":"ContainerDied","Data":"75b288f79920a57cf39dd57ed532721dab0f87ebf88519013a2061e8f811e374"} Dec 09 17:27:53 crc kubenswrapper[4954]: I1209 17:27:53.150305 4954 generic.go:334] "Generic (PLEG): container finished" podID="9653f9d1-0297-458e-bfb6-b6801ef96c3c" containerID="9ce2957549eff70070e4dc18369586f0756e0c26b3c09557904e5f37ab26085c" exitCode=0 Dec 09 17:27:53 crc kubenswrapper[4954]: I1209 17:27:53.150698 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r98b8" event={"ID":"9653f9d1-0297-458e-bfb6-b6801ef96c3c","Type":"ContainerDied","Data":"9ce2957549eff70070e4dc18369586f0756e0c26b3c09557904e5f37ab26085c"} Dec 09 17:27:54 crc kubenswrapper[4954]: I1209 17:27:54.891662 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-p5ndh" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.011887 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-config-data\") pod \"25667d18-6643-4344-9ac3-e3a4f1af6957\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.012164 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-combined-ca-bundle\") pod \"25667d18-6643-4344-9ac3-e3a4f1af6957\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.012336 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wp75\" (UniqueName: \"kubernetes.io/projected/25667d18-6643-4344-9ac3-e3a4f1af6957-kube-api-access-5wp75\") pod \"25667d18-6643-4344-9ac3-e3a4f1af6957\" (UID: \"25667d18-6643-4344-9ac3-e3a4f1af6957\") " Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.018949 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25667d18-6643-4344-9ac3-e3a4f1af6957-kube-api-access-5wp75" (OuterVolumeSpecName: "kube-api-access-5wp75") pod "25667d18-6643-4344-9ac3-e3a4f1af6957" (UID: "25667d18-6643-4344-9ac3-e3a4f1af6957"). InnerVolumeSpecName "kube-api-access-5wp75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.048497 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25667d18-6643-4344-9ac3-e3a4f1af6957" (UID: "25667d18-6643-4344-9ac3-e3a4f1af6957"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.109225 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-config-data" (OuterVolumeSpecName: "config-data") pod "25667d18-6643-4344-9ac3-e3a4f1af6957" (UID: "25667d18-6643-4344-9ac3-e3a4f1af6957"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.115160 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wp75\" (UniqueName: \"kubernetes.io/projected/25667d18-6643-4344-9ac3-e3a4f1af6957-kube-api-access-5wp75\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.115192 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.115204 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25667d18-6643-4344-9ac3-e3a4f1af6957-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.186798 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-p5ndh" event={"ID":"25667d18-6643-4344-9ac3-e3a4f1af6957","Type":"ContainerDied","Data":"ea4239307ccedc6fc923097af10c147bf617eb10495c775c3dda6207306a296c"} Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.186845 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea4239307ccedc6fc923097af10c147bf617eb10495c775c3dda6207306a296c" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.186888 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-p5ndh" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.413851 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.523924 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9653f9d1-0297-458e-bfb6-b6801ef96c3c-etc-machine-id\") pod \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.524068 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9653f9d1-0297-458e-bfb6-b6801ef96c3c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9653f9d1-0297-458e-bfb6-b6801ef96c3c" (UID: "9653f9d1-0297-458e-bfb6-b6801ef96c3c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.524123 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4kr6\" (UniqueName: \"kubernetes.io/projected/9653f9d1-0297-458e-bfb6-b6801ef96c3c-kube-api-access-l4kr6\") pod \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.524279 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-config-data\") pod \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.524309 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-scripts\") pod \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.524349 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-combined-ca-bundle\") pod \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.524458 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-db-sync-config-data\") pod \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\" (UID: \"9653f9d1-0297-458e-bfb6-b6801ef96c3c\") " Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.524938 4954 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9653f9d1-0297-458e-bfb6-b6801ef96c3c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.530425 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9653f9d1-0297-458e-bfb6-b6801ef96c3c-kube-api-access-l4kr6" (OuterVolumeSpecName: "kube-api-access-l4kr6") pod "9653f9d1-0297-458e-bfb6-b6801ef96c3c" (UID: "9653f9d1-0297-458e-bfb6-b6801ef96c3c"). InnerVolumeSpecName "kube-api-access-l4kr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.530713 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9653f9d1-0297-458e-bfb6-b6801ef96c3c" (UID: "9653f9d1-0297-458e-bfb6-b6801ef96c3c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.532109 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-scripts" (OuterVolumeSpecName: "scripts") pod "9653f9d1-0297-458e-bfb6-b6801ef96c3c" (UID: "9653f9d1-0297-458e-bfb6-b6801ef96c3c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.558269 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9653f9d1-0297-458e-bfb6-b6801ef96c3c" (UID: "9653f9d1-0297-458e-bfb6-b6801ef96c3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.588018 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-config-data" (OuterVolumeSpecName: "config-data") pod "9653f9d1-0297-458e-bfb6-b6801ef96c3c" (UID: "9653f9d1-0297-458e-bfb6-b6801ef96c3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.627553 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.627878 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.627892 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.627909 4954 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9653f9d1-0297-458e-bfb6-b6801ef96c3c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.627921 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4kr6\" (UniqueName: \"kubernetes.io/projected/9653f9d1-0297-458e-bfb6-b6801ef96c3c-kube-api-access-l4kr6\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.659109 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.731966 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-77x96"] Dec 09 17:27:55 crc kubenswrapper[4954]: I1209 17:27:55.732295 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" podUID="83aa4b70-c6f9-413f-9d19-65f2b9701ecf" containerName="dnsmasq-dns" containerID="cri-o://aaaa514a40281435d4c5223f2f03d3f099b06add2067aa09d4d3b9640768888f" gracePeriod=10 Dec 09 17:27:55 crc kubenswrapper[4954]: E1209 17:27:55.832108 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.204339 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r98b8" event={"ID":"9653f9d1-0297-458e-bfb6-b6801ef96c3c","Type":"ContainerDied","Data":"7e5d8deda681bc17d1ebb32ba6ecf702f9d9a41da2de897794ea7fb2165a106d"} Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.204761 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e5d8deda681bc17d1ebb32ba6ecf702f9d9a41da2de897794ea7fb2165a106d" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.204712 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r98b8" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.213181 4954 generic.go:334] "Generic (PLEG): container finished" podID="83aa4b70-c6f9-413f-9d19-65f2b9701ecf" containerID="aaaa514a40281435d4c5223f2f03d3f099b06add2067aa09d4d3b9640768888f" exitCode=0 Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.213266 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" event={"ID":"83aa4b70-c6f9-413f-9d19-65f2b9701ecf","Type":"ContainerDied","Data":"aaaa514a40281435d4c5223f2f03d3f099b06add2067aa09d4d3b9640768888f"} Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.239993 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"565df52a-04b4-463b-a1b1-59eee6d94ea5","Type":"ContainerStarted","Data":"3beaa2793d2d123bac063806a084314f95d531c8a357d2dcf571129db1d03073"} Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.240260 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="ceilometer-notification-agent" containerID="cri-o://69758858f95933f87c890e8522b414326890401c5401220544f9a87b466ba635" gracePeriod=30 Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.240416 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.241013 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="proxy-httpd" containerID="cri-o://3beaa2793d2d123bac063806a084314f95d531c8a357d2dcf571129db1d03073" gracePeriod=30 Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.241087 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="sg-core" containerID="cri-o://618ff916e034b8d1cc36791523390bc905f827eb4053d62b5486e49cb605cfc4" gracePeriod=30 Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.276801 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6dfdb467d-r2rqf" event={"ID":"653b2aa3-1e55-42ae-9284-51d651461a15","Type":"ContainerStarted","Data":"4b58a75379599700147bfd37376dbeab5f6e959ad6063ea27e16d5b2849b1ab8"} Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.278654 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.279020 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.387668 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6dfdb467d-r2rqf" podStartSLOduration=8.387642856 podStartE2EDuration="8.387642856s" podCreationTimestamp="2025-12-09 17:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:27:56.311714759 +0000 UTC m=+1872.699888589" watchObservedRunningTime="2025-12-09 17:27:56.387642856 +0000 UTC m=+1872.775816666" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.446733 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.580481 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-config\") pod \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.580580 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mvwq\" (UniqueName: \"kubernetes.io/projected/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-kube-api-access-9mvwq\") pod \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.580711 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-svc\") pod \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.580852 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-nb\") pod \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.580891 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-sb\") pod \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.580922 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-swift-storage-0\") pod \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\" (UID: \"83aa4b70-c6f9-413f-9d19-65f2b9701ecf\") " Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.636762 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-kube-api-access-9mvwq" (OuterVolumeSpecName: "kube-api-access-9mvwq") pod "83aa4b70-c6f9-413f-9d19-65f2b9701ecf" (UID: "83aa4b70-c6f9-413f-9d19-65f2b9701ecf"). InnerVolumeSpecName "kube-api-access-9mvwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.683582 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mvwq\" (UniqueName: \"kubernetes.io/projected/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-kube-api-access-9mvwq\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.713544 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "83aa4b70-c6f9-413f-9d19-65f2b9701ecf" (UID: "83aa4b70-c6f9-413f-9d19-65f2b9701ecf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.719130 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "83aa4b70-c6f9-413f-9d19-65f2b9701ecf" (UID: "83aa4b70-c6f9-413f-9d19-65f2b9701ecf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.733751 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 17:27:56 crc kubenswrapper[4954]: E1209 17:27:56.734356 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25667d18-6643-4344-9ac3-e3a4f1af6957" containerName="heat-db-sync" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.734376 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="25667d18-6643-4344-9ac3-e3a4f1af6957" containerName="heat-db-sync" Dec 09 17:27:56 crc kubenswrapper[4954]: E1209 17:27:56.734399 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9653f9d1-0297-458e-bfb6-b6801ef96c3c" containerName="cinder-db-sync" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.734409 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9653f9d1-0297-458e-bfb6-b6801ef96c3c" containerName="cinder-db-sync" Dec 09 17:27:56 crc kubenswrapper[4954]: E1209 17:27:56.734428 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83aa4b70-c6f9-413f-9d19-65f2b9701ecf" containerName="init" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.734437 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="83aa4b70-c6f9-413f-9d19-65f2b9701ecf" containerName="init" Dec 09 17:27:56 crc kubenswrapper[4954]: E1209 17:27:56.734474 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83aa4b70-c6f9-413f-9d19-65f2b9701ecf" containerName="dnsmasq-dns" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.734482 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="83aa4b70-c6f9-413f-9d19-65f2b9701ecf" containerName="dnsmasq-dns" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.734718 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="83aa4b70-c6f9-413f-9d19-65f2b9701ecf" containerName="dnsmasq-dns" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.734738 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="9653f9d1-0297-458e-bfb6-b6801ef96c3c" containerName="cinder-db-sync" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.734745 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="25667d18-6643-4344-9ac3-e3a4f1af6957" containerName="heat-db-sync" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.742486 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.744918 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.763315 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.763653 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.763793 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-npkrm" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.763957 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.776744 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "83aa4b70-c6f9-413f-9d19-65f2b9701ecf" (UID: "83aa4b70-c6f9-413f-9d19-65f2b9701ecf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.797154 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-scripts\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.797305 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.797374 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.797559 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2wl4\" (UniqueName: \"kubernetes.io/projected/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-kube-api-access-n2wl4\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.797665 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.797960 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.798162 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.798181 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.798192 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.806708 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-config" (OuterVolumeSpecName: "config") pod "83aa4b70-c6f9-413f-9d19-65f2b9701ecf" (UID: "83aa4b70-c6f9-413f-9d19-65f2b9701ecf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.843178 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "83aa4b70-c6f9-413f-9d19-65f2b9701ecf" (UID: "83aa4b70-c6f9-413f-9d19-65f2b9701ecf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.866950 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lffjc"] Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.869239 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.871837 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lffjc"] Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.901489 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.901541 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.901581 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.901670 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2wl4\" (UniqueName: \"kubernetes.io/projected/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-kube-api-access-n2wl4\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.901691 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.901712 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.902770 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-config\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.902891 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.903145 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.903182 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.903255 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwdhw\" (UniqueName: \"kubernetes.io/projected/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-kube-api-access-dwdhw\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.903289 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.903353 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-scripts\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.903498 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.903519 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83aa4b70-c6f9-413f-9d19-65f2b9701ecf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.918202 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.919322 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.926793 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.943370 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-scripts\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:56 crc kubenswrapper[4954]: I1209 17:27:56.945209 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2wl4\" (UniqueName: \"kubernetes.io/projected/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-kube-api-access-n2wl4\") pod \"cinder-scheduler-0\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " pod="openstack/cinder-scheduler-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.035764 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.036021 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwdhw\" (UniqueName: \"kubernetes.io/projected/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-kube-api-access-dwdhw\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.036103 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.036301 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.036537 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-config\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.036630 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.040186 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.041318 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.042089 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.043077 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-config\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.053329 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.094637 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwdhw\" (UniqueName: \"kubernetes.io/projected/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-kube-api-access-dwdhw\") pod \"dnsmasq-dns-5c9776ccc5-lffjc\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.102950 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.108416 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.111914 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.115622 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.162145 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.250633 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data-custom\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.250694 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.250815 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhwfx\" (UniqueName: \"kubernetes.io/projected/d2a08111-90cf-4b01-9427-0dc63f2c5983-kube-api-access-hhwfx\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.250844 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2a08111-90cf-4b01-9427-0dc63f2c5983-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.250938 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.250995 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-scripts\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.251025 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2a08111-90cf-4b01-9427-0dc63f2c5983-logs\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.290523 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" event={"ID":"83aa4b70-c6f9-413f-9d19-65f2b9701ecf","Type":"ContainerDied","Data":"327b98bd762c21f81bf271a885bea9eccb06aa0d15565894ecf17d60f11b6d50"} Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.291142 4954 scope.go:117] "RemoveContainer" containerID="aaaa514a40281435d4c5223f2f03d3f099b06add2067aa09d4d3b9640768888f" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.290807 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-77x96" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.295245 4954 generic.go:334] "Generic (PLEG): container finished" podID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerID="3beaa2793d2d123bac063806a084314f95d531c8a357d2dcf571129db1d03073" exitCode=0 Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.295279 4954 generic.go:334] "Generic (PLEG): container finished" podID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerID="618ff916e034b8d1cc36791523390bc905f827eb4053d62b5486e49cb605cfc4" exitCode=2 Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.295334 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"565df52a-04b4-463b-a1b1-59eee6d94ea5","Type":"ContainerDied","Data":"3beaa2793d2d123bac063806a084314f95d531c8a357d2dcf571129db1d03073"} Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.295412 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"565df52a-04b4-463b-a1b1-59eee6d94ea5","Type":"ContainerDied","Data":"618ff916e034b8d1cc36791523390bc905f827eb4053d62b5486e49cb605cfc4"} Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.339050 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-77x96"] Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.353965 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data-custom\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.354043 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.355656 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-77x96"] Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.355801 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhwfx\" (UniqueName: \"kubernetes.io/projected/d2a08111-90cf-4b01-9427-0dc63f2c5983-kube-api-access-hhwfx\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.356221 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2a08111-90cf-4b01-9427-0dc63f2c5983-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.356457 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.356621 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-scripts\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.356676 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2a08111-90cf-4b01-9427-0dc63f2c5983-logs\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.357192 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2a08111-90cf-4b01-9427-0dc63f2c5983-logs\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.357260 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2a08111-90cf-4b01-9427-0dc63f2c5983-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.361513 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data-custom\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.362717 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-scripts\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.363473 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.365164 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.376384 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhwfx\" (UniqueName: \"kubernetes.io/projected/d2a08111-90cf-4b01-9427-0dc63f2c5983-kube-api-access-hhwfx\") pod \"cinder-api-0\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " pod="openstack/cinder-api-0" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.389625 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:27:57 crc kubenswrapper[4954]: I1209 17:27:57.466764 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 17:27:58 crc kubenswrapper[4954]: I1209 17:27:58.143779 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83aa4b70-c6f9-413f-9d19-65f2b9701ecf" path="/var/lib/kubelet/pods/83aa4b70-c6f9-413f-9d19-65f2b9701ecf/volumes" Dec 09 17:27:58 crc kubenswrapper[4954]: I1209 17:27:58.319419 4954 generic.go:334] "Generic (PLEG): container finished" podID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerID="69758858f95933f87c890e8522b414326890401c5401220544f9a87b466ba635" exitCode=0 Dec 09 17:27:58 crc kubenswrapper[4954]: I1209 17:27:58.319518 4954 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 17:27:58 crc kubenswrapper[4954]: I1209 17:27:58.320510 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"565df52a-04b4-463b-a1b1-59eee6d94ea5","Type":"ContainerDied","Data":"69758858f95933f87c890e8522b414326890401c5401220544f9a87b466ba635"} Dec 09 17:27:58 crc kubenswrapper[4954]: I1209 17:27:58.360415 4954 scope.go:117] "RemoveContainer" containerID="a8e4b24d4b0576a1066c670e11eaaaff03001241fb677f63cdc51997f76a1560" Dec 09 17:27:58 crc kubenswrapper[4954]: I1209 17:27:58.717548 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:58 crc kubenswrapper[4954]: I1209 17:27:58.964658 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:27:58 crc kubenswrapper[4954]: I1209 17:27:58.973777 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.120869 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.121094 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 17:27:59 crc kubenswrapper[4954]: E1209 17:27:59.121210 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.131431 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-combined-ca-bundle\") pod \"565df52a-04b4-463b-a1b1-59eee6d94ea5\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.131519 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-scripts\") pod \"565df52a-04b4-463b-a1b1-59eee6d94ea5\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.131658 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd225\" (UniqueName: \"kubernetes.io/projected/565df52a-04b4-463b-a1b1-59eee6d94ea5-kube-api-access-vd225\") pod \"565df52a-04b4-463b-a1b1-59eee6d94ea5\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.131711 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-log-httpd\") pod \"565df52a-04b4-463b-a1b1-59eee6d94ea5\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.131878 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-sg-core-conf-yaml\") pod \"565df52a-04b4-463b-a1b1-59eee6d94ea5\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.131912 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-run-httpd\") pod \"565df52a-04b4-463b-a1b1-59eee6d94ea5\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.131994 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-config-data\") pod \"565df52a-04b4-463b-a1b1-59eee6d94ea5\" (UID: \"565df52a-04b4-463b-a1b1-59eee6d94ea5\") " Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.134510 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "565df52a-04b4-463b-a1b1-59eee6d94ea5" (UID: "565df52a-04b4-463b-a1b1-59eee6d94ea5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.135036 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "565df52a-04b4-463b-a1b1-59eee6d94ea5" (UID: "565df52a-04b4-463b-a1b1-59eee6d94ea5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.235041 4954 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.235732 4954 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/565df52a-04b4-463b-a1b1-59eee6d94ea5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.335581 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"565df52a-04b4-463b-a1b1-59eee6d94ea5","Type":"ContainerDied","Data":"68ca007837b4e705a4238cfe1f271e417faacdbf93bf0f4fcdc0b934207f2560"} Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.335671 4954 scope.go:117] "RemoveContainer" containerID="3beaa2793d2d123bac063806a084314f95d531c8a357d2dcf571129db1d03073" Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.337469 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.342266 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6c7b7c8-3be0-4271-8ecd-f5db01209de7","Type":"ContainerStarted","Data":"e50c5d15ad1b17cca20383c9d5ff2d88a9495cf4b52296ff93eef890289b6225"} Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.362841 4954 scope.go:117] "RemoveContainer" containerID="618ff916e034b8d1cc36791523390bc905f827eb4053d62b5486e49cb605cfc4" Dec 09 17:27:59 crc kubenswrapper[4954]: I1209 17:27:59.383847 4954 scope.go:117] "RemoveContainer" containerID="69758858f95933f87c890e8522b414326890401c5401220544f9a87b466ba635" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.074943 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-scripts" (OuterVolumeSpecName: "scripts") pod "565df52a-04b4-463b-a1b1-59eee6d94ea5" (UID: "565df52a-04b4-463b-a1b1-59eee6d94ea5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.083910 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565df52a-04b4-463b-a1b1-59eee6d94ea5-kube-api-access-vd225" (OuterVolumeSpecName: "kube-api-access-vd225") pod "565df52a-04b4-463b-a1b1-59eee6d94ea5" (UID: "565df52a-04b4-463b-a1b1-59eee6d94ea5"). InnerVolumeSpecName "kube-api-access-vd225". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.099302 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "565df52a-04b4-463b-a1b1-59eee6d94ea5" (UID: "565df52a-04b4-463b-a1b1-59eee6d94ea5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.133125 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "565df52a-04b4-463b-a1b1-59eee6d94ea5" (UID: "565df52a-04b4-463b-a1b1-59eee6d94ea5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.158668 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.158701 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.158716 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd225\" (UniqueName: \"kubernetes.io/projected/565df52a-04b4-463b-a1b1-59eee6d94ea5-kube-api-access-vd225\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.158733 4954 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.254681 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lffjc"] Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.255098 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.332867 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-config-data" (OuterVolumeSpecName: "config-data") pod "565df52a-04b4-463b-a1b1-59eee6d94ea5" (UID: "565df52a-04b4-463b-a1b1-59eee6d94ea5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.394850 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" event={"ID":"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be","Type":"ContainerStarted","Data":"1c81aad3430743b8ebe79fcc0b892cf6f12b0845bc9d379cabbf5378db4647b7"} Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.397512 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/565df52a-04b4-463b-a1b1-59eee6d94ea5-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.404623 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d2a08111-90cf-4b01-9427-0dc63f2c5983","Type":"ContainerStarted","Data":"1d5763e1a19fc7c7c846498f5e664275fbb38d5bd5e9affcc8ee165753173e25"} Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.640717 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.685050 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.694791 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.709012 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:00 crc kubenswrapper[4954]: E1209 17:28:00.710024 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="ceilometer-notification-agent" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.710046 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="ceilometer-notification-agent" Dec 09 17:28:00 crc kubenswrapper[4954]: E1209 17:28:00.710068 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="sg-core" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.710076 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="sg-core" Dec 09 17:28:00 crc kubenswrapper[4954]: E1209 17:28:00.710118 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="proxy-httpd" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.710128 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="proxy-httpd" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.710350 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="sg-core" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.710366 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="proxy-httpd" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.710396 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" containerName="ceilometer-notification-agent" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.712819 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.716081 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.717546 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.728715 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.801586 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.808382 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-run-httpd\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.808456 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.808486 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhghg\" (UniqueName: \"kubernetes.io/projected/2224489a-740d-4225-bd71-f21cca0586f3-kube-api-access-nhghg\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.808544 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-scripts\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.808566 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.808618 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-config-data\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.808633 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-log-httpd\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.910823 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-config-data\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.910890 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-log-httpd\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.911082 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-run-httpd\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.911154 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.911214 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhghg\" (UniqueName: \"kubernetes.io/projected/2224489a-740d-4225-bd71-f21cca0586f3-kube-api-access-nhghg\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.911332 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-scripts\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.911384 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.913452 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-log-httpd\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.914890 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-run-httpd\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.924227 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.930120 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-scripts\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.931198 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.933848 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-config-data\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:00 crc kubenswrapper[4954]: I1209 17:28:00.934880 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhghg\" (UniqueName: \"kubernetes.io/projected/2224489a-740d-4225-bd71-f21cca0586f3-kube-api-access-nhghg\") pod \"ceilometer-0\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " pod="openstack/ceilometer-0" Dec 09 17:28:01 crc kubenswrapper[4954]: I1209 17:28:01.133234 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:28:01 crc kubenswrapper[4954]: I1209 17:28:01.435573 4954 generic.go:334] "Generic (PLEG): container finished" podID="d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" containerID="e3a76c6136ded675b2709fb350f57424cb4b0ab3524a287e16f7ffe4c028bf9a" exitCode=0 Dec 09 17:28:01 crc kubenswrapper[4954]: I1209 17:28:01.436060 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" event={"ID":"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be","Type":"ContainerDied","Data":"e3a76c6136ded675b2709fb350f57424cb4b0ab3524a287e16f7ffe4c028bf9a"} Dec 09 17:28:01 crc kubenswrapper[4954]: I1209 17:28:01.782119 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:28:01 crc kubenswrapper[4954]: I1209 17:28:01.909628 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:02 crc kubenswrapper[4954]: I1209 17:28:02.148368 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="565df52a-04b4-463b-a1b1-59eee6d94ea5" path="/var/lib/kubelet/pods/565df52a-04b4-463b-a1b1-59eee6d94ea5/volumes" Dec 09 17:28:02 crc kubenswrapper[4954]: I1209 17:28:02.456404 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2224489a-740d-4225-bd71-f21cca0586f3","Type":"ContainerStarted","Data":"b144a2f4766e0ec40f31400d3e224f3bb179dca3c5733f18d89e838578cb0c90"} Dec 09 17:28:02 crc kubenswrapper[4954]: I1209 17:28:02.459834 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" event={"ID":"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be","Type":"ContainerStarted","Data":"f14e6b3ced7be04a32d22b82a3f3acd5a88b7b5a16f4970ad96ca1e6bb5ca326"} Dec 09 17:28:02 crc kubenswrapper[4954]: I1209 17:28:02.460775 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:28:02 crc kubenswrapper[4954]: I1209 17:28:02.464902 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d2a08111-90cf-4b01-9427-0dc63f2c5983","Type":"ContainerStarted","Data":"2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae"} Dec 09 17:28:02 crc kubenswrapper[4954]: I1209 17:28:02.468251 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6c7b7c8-3be0-4271-8ecd-f5db01209de7","Type":"ContainerStarted","Data":"4a0c6b84827b9126618fe1522e0df3091d401ee36921bbefbc385b4ffa902807"} Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.433234 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6dfdb467d-r2rqf" Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.469185 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" podStartSLOduration=7.469150506 podStartE2EDuration="7.469150506s" podCreationTimestamp="2025-12-09 17:27:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:02.495739553 +0000 UTC m=+1878.883913403" watchObservedRunningTime="2025-12-09 17:28:03.469150506 +0000 UTC m=+1879.857324326" Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.481981 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d2a08111-90cf-4b01-9427-0dc63f2c5983","Type":"ContainerStarted","Data":"b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab"} Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.482134 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d2a08111-90cf-4b01-9427-0dc63f2c5983" containerName="cinder-api-log" containerID="cri-o://2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae" gracePeriod=30 Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.482202 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.482258 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d2a08111-90cf-4b01-9427-0dc63f2c5983" containerName="cinder-api" containerID="cri-o://b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab" gracePeriod=30 Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.502854 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6c7b7c8-3be0-4271-8ecd-f5db01209de7","Type":"ContainerStarted","Data":"de8b7c88c529ea3160f4ec26701963ce9790d28d8c82fc745e15191475a66b2d"} Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.522123 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2224489a-740d-4225-bd71-f21cca0586f3","Type":"ContainerStarted","Data":"eec4ad66049242ba5deee324cf61411d2975c4d953a0269250822af455084537"} Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.540143 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-f98894b5d-v8d8l"] Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.540446 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-f98894b5d-v8d8l" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerName="barbican-api-log" containerID="cri-o://f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38" gracePeriod=30 Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.540642 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-f98894b5d-v8d8l" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerName="barbican-api" containerID="cri-o://1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea" gracePeriod=30 Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.581200 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.581178613 podStartE2EDuration="7.581178613s" podCreationTimestamp="2025-12-09 17:27:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:03.54562859 +0000 UTC m=+1879.933802420" watchObservedRunningTime="2025-12-09 17:28:03.581178613 +0000 UTC m=+1879.969352433" Dec 09 17:28:03 crc kubenswrapper[4954]: I1209 17:28:03.607851 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.291430617 podStartE2EDuration="7.607828098s" podCreationTimestamp="2025-12-09 17:27:56 +0000 UTC" firstStartedPulling="2025-12-09 17:27:59.079792235 +0000 UTC m=+1875.467966055" lastFinishedPulling="2025-12-09 17:28:00.396189716 +0000 UTC m=+1876.784363536" observedRunningTime="2025-12-09 17:28:03.579380697 +0000 UTC m=+1879.967554517" watchObservedRunningTime="2025-12-09 17:28:03.607828098 +0000 UTC m=+1879.996001918" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.347076 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.430782 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-9855f74b5-lqsx6" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.526789 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data\") pod \"d2a08111-90cf-4b01-9427-0dc63f2c5983\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.526909 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2a08111-90cf-4b01-9427-0dc63f2c5983-logs\") pod \"d2a08111-90cf-4b01-9427-0dc63f2c5983\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.526964 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhwfx\" (UniqueName: \"kubernetes.io/projected/d2a08111-90cf-4b01-9427-0dc63f2c5983-kube-api-access-hhwfx\") pod \"d2a08111-90cf-4b01-9427-0dc63f2c5983\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.527015 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-combined-ca-bundle\") pod \"d2a08111-90cf-4b01-9427-0dc63f2c5983\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.527137 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-scripts\") pod \"d2a08111-90cf-4b01-9427-0dc63f2c5983\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.527250 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2a08111-90cf-4b01-9427-0dc63f2c5983-etc-machine-id\") pod \"d2a08111-90cf-4b01-9427-0dc63f2c5983\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.527307 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data-custom\") pod \"d2a08111-90cf-4b01-9427-0dc63f2c5983\" (UID: \"d2a08111-90cf-4b01-9427-0dc63f2c5983\") " Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.536025 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2a08111-90cf-4b01-9427-0dc63f2c5983-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d2a08111-90cf-4b01-9427-0dc63f2c5983" (UID: "d2a08111-90cf-4b01-9427-0dc63f2c5983"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.536343 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2a08111-90cf-4b01-9427-0dc63f2c5983-logs" (OuterVolumeSpecName: "logs") pod "d2a08111-90cf-4b01-9427-0dc63f2c5983" (UID: "d2a08111-90cf-4b01-9427-0dc63f2c5983"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.542945 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d2a08111-90cf-4b01-9427-0dc63f2c5983" (UID: "d2a08111-90cf-4b01-9427-0dc63f2c5983"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.548445 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-scripts" (OuterVolumeSpecName: "scripts") pod "d2a08111-90cf-4b01-9427-0dc63f2c5983" (UID: "d2a08111-90cf-4b01-9427-0dc63f2c5983"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.550110 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2224489a-740d-4225-bd71-f21cca0586f3","Type":"ContainerStarted","Data":"3b24806e1afa23c91c40dcd2b739d3c623595b8232defe75355958cb931fa427"} Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.550141 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2a08111-90cf-4b01-9427-0dc63f2c5983-kube-api-access-hhwfx" (OuterVolumeSpecName: "kube-api-access-hhwfx") pod "d2a08111-90cf-4b01-9427-0dc63f2c5983" (UID: "d2a08111-90cf-4b01-9427-0dc63f2c5983"). InnerVolumeSpecName "kube-api-access-hhwfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.559486 4954 generic.go:334] "Generic (PLEG): container finished" podID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerID="f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38" exitCode=143 Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.560156 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f98894b5d-v8d8l" event={"ID":"9b5721ed-4849-42c5-90c8-5524bcd5d9e1","Type":"ContainerDied","Data":"f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38"} Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.578839 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-76bd5dc95b-mwznb"] Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.579132 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-76bd5dc95b-mwznb" podUID="284a949a-68ec-4470-a956-6e0a396675e0" containerName="neutron-api" containerID="cri-o://58892b864944f9d101cc233d21c8b6e6f2e20e90c186a670c72216c198b1c92e" gracePeriod=30 Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.579263 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-76bd5dc95b-mwznb" podUID="284a949a-68ec-4470-a956-6e0a396675e0" containerName="neutron-httpd" containerID="cri-o://67698af236753d0e979ade240e6b3c96329e70ec61deb85d7206df9b666141d6" gracePeriod=30 Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.594260 4954 generic.go:334] "Generic (PLEG): container finished" podID="d2a08111-90cf-4b01-9427-0dc63f2c5983" containerID="b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab" exitCode=0 Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.595124 4954 generic.go:334] "Generic (PLEG): container finished" podID="d2a08111-90cf-4b01-9427-0dc63f2c5983" containerID="2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae" exitCode=143 Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.594466 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.594327 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d2a08111-90cf-4b01-9427-0dc63f2c5983","Type":"ContainerDied","Data":"b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab"} Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.602692 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d2a08111-90cf-4b01-9427-0dc63f2c5983","Type":"ContainerDied","Data":"2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae"} Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.602723 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d2a08111-90cf-4b01-9427-0dc63f2c5983","Type":"ContainerDied","Data":"1d5763e1a19fc7c7c846498f5e664275fbb38d5bd5e9affcc8ee165753173e25"} Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.602747 4954 scope.go:117] "RemoveContainer" containerID="b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.616500 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2a08111-90cf-4b01-9427-0dc63f2c5983" (UID: "d2a08111-90cf-4b01-9427-0dc63f2c5983"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.639416 4954 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.639466 4954 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2a08111-90cf-4b01-9427-0dc63f2c5983-logs\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.639479 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhwfx\" (UniqueName: \"kubernetes.io/projected/d2a08111-90cf-4b01-9427-0dc63f2c5983-kube-api-access-hhwfx\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.639487 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.639498 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.639506 4954 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2a08111-90cf-4b01-9427-0dc63f2c5983-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.698867 4954 scope.go:117] "RemoveContainer" containerID="2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.704097 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data" (OuterVolumeSpecName: "config-data") pod "d2a08111-90cf-4b01-9427-0dc63f2c5983" (UID: "d2a08111-90cf-4b01-9427-0dc63f2c5983"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.743066 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2a08111-90cf-4b01-9427-0dc63f2c5983-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.747432 4954 scope.go:117] "RemoveContainer" containerID="b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab" Dec 09 17:28:04 crc kubenswrapper[4954]: E1209 17:28:04.748685 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab\": container with ID starting with b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab not found: ID does not exist" containerID="b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.748735 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab"} err="failed to get container status \"b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab\": rpc error: code = NotFound desc = could not find container \"b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab\": container with ID starting with b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab not found: ID does not exist" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.748766 4954 scope.go:117] "RemoveContainer" containerID="2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae" Dec 09 17:28:04 crc kubenswrapper[4954]: E1209 17:28:04.749160 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae\": container with ID starting with 2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae not found: ID does not exist" containerID="2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.749203 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae"} err="failed to get container status \"2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae\": rpc error: code = NotFound desc = could not find container \"2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae\": container with ID starting with 2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae not found: ID does not exist" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.749224 4954 scope.go:117] "RemoveContainer" containerID="b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.767053 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab"} err="failed to get container status \"b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab\": rpc error: code = NotFound desc = could not find container \"b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab\": container with ID starting with b1d4cd25009ae1c52962f3f1a789e70a4a909614ccd18e1ab491b6f35e59e9ab not found: ID does not exist" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.767125 4954 scope.go:117] "RemoveContainer" containerID="2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.768099 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae"} err="failed to get container status \"2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae\": rpc error: code = NotFound desc = could not find container \"2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae\": container with ID starting with 2e9e302c88bd5601e797a8bf62a48b573224fb3a30be996aa6059ce35ac329ae not found: ID does not exist" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.947447 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.968045 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.984931 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 09 17:28:04 crc kubenswrapper[4954]: E1209 17:28:04.985700 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a08111-90cf-4b01-9427-0dc63f2c5983" containerName="cinder-api" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.985725 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a08111-90cf-4b01-9427-0dc63f2c5983" containerName="cinder-api" Dec 09 17:28:04 crc kubenswrapper[4954]: E1209 17:28:04.985760 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a08111-90cf-4b01-9427-0dc63f2c5983" containerName="cinder-api-log" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.985769 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a08111-90cf-4b01-9427-0dc63f2c5983" containerName="cinder-api-log" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.986020 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2a08111-90cf-4b01-9427-0dc63f2c5983" containerName="cinder-api" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.986043 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2a08111-90cf-4b01-9427-0dc63f2c5983" containerName="cinder-api-log" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.987376 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.992143 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.995092 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 09 17:28:04 crc kubenswrapper[4954]: I1209 17:28:04.995209 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.000208 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.155840 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-config-data\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.155920 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8078c26-0c70-430b-b3d8-6f56d236657c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.155975 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-scripts\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.156002 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.156067 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.156321 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8078c26-0c70-430b-b3d8-6f56d236657c-logs\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.156436 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfbbj\" (UniqueName: \"kubernetes.io/projected/a8078c26-0c70-430b-b3d8-6f56d236657c-kube-api-access-mfbbj\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.156546 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.156662 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-config-data-custom\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.259498 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-config-data-custom\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.260133 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-config-data\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.261125 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8078c26-0c70-430b-b3d8-6f56d236657c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.261358 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-scripts\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.261443 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8078c26-0c70-430b-b3d8-6f56d236657c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.261456 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.261649 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.261887 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8078c26-0c70-430b-b3d8-6f56d236657c-logs\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.262160 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfbbj\" (UniqueName: \"kubernetes.io/projected/a8078c26-0c70-430b-b3d8-6f56d236657c-kube-api-access-mfbbj\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.262288 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.263295 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8078c26-0c70-430b-b3d8-6f56d236657c-logs\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.264340 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-config-data-custom\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.265236 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-config-data\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.266222 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-scripts\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.265794 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.269825 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.271552 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8078c26-0c70-430b-b3d8-6f56d236657c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.292528 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfbbj\" (UniqueName: \"kubernetes.io/projected/a8078c26-0c70-430b-b3d8-6f56d236657c-kube-api-access-mfbbj\") pod \"cinder-api-0\" (UID: \"a8078c26-0c70-430b-b3d8-6f56d236657c\") " pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.310239 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.620152 4954 generic.go:334] "Generic (PLEG): container finished" podID="284a949a-68ec-4470-a956-6e0a396675e0" containerID="67698af236753d0e979ade240e6b3c96329e70ec61deb85d7206df9b666141d6" exitCode=0 Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.620218 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76bd5dc95b-mwznb" event={"ID":"284a949a-68ec-4470-a956-6e0a396675e0","Type":"ContainerDied","Data":"67698af236753d0e979ade240e6b3c96329e70ec61deb85d7206df9b666141d6"} Dec 09 17:28:05 crc kubenswrapper[4954]: W1209 17:28:05.918512 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8078c26_0c70_430b_b3d8_6f56d236657c.slice/crio-f92a3c886a86b97f642d9e687d92435d3fa67b66f51ff77e2e55221e5f9d70f1 WatchSource:0}: Error finding container f92a3c886a86b97f642d9e687d92435d3fa67b66f51ff77e2e55221e5f9d70f1: Status 404 returned error can't find the container with id f92a3c886a86b97f642d9e687d92435d3fa67b66f51ff77e2e55221e5f9d70f1 Dec 09 17:28:05 crc kubenswrapper[4954]: I1209 17:28:05.926108 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 17:28:06 crc kubenswrapper[4954]: I1209 17:28:06.134091 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2a08111-90cf-4b01-9427-0dc63f2c5983" path="/var/lib/kubelet/pods/d2a08111-90cf-4b01-9427-0dc63f2c5983/volumes" Dec 09 17:28:06 crc kubenswrapper[4954]: I1209 17:28:06.634413 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2224489a-740d-4225-bd71-f21cca0586f3","Type":"ContainerStarted","Data":"3f99c83fa80b3b9f05baf5a6d89cf0b52b220d2d4cecfbef8f7c1cbbe2e6eab0"} Dec 09 17:28:06 crc kubenswrapper[4954]: I1209 17:28:06.636762 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a8078c26-0c70-430b-b3d8-6f56d236657c","Type":"ContainerStarted","Data":"f38d00752e82940cbb911e22cd0b722a8bd3946622af46d9c1f95852d61d45fd"} Dec 09 17:28:06 crc kubenswrapper[4954]: I1209 17:28:06.636804 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a8078c26-0c70-430b-b3d8-6f56d236657c","Type":"ContainerStarted","Data":"f92a3c886a86b97f642d9e687d92435d3fa67b66f51ff77e2e55221e5f9d70f1"} Dec 09 17:28:06 crc kubenswrapper[4954]: I1209 17:28:06.737857 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f98894b5d-v8d8l" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.190:9311/healthcheck\": read tcp 10.217.0.2:46006->10.217.0.190:9311: read: connection reset by peer" Dec 09 17:28:06 crc kubenswrapper[4954]: I1209 17:28:06.738500 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f98894b5d-v8d8l" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.190:9311/healthcheck\": read tcp 10.217.0.2:46014->10.217.0.190:9311: read: connection reset by peer" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.116316 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.309101 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.393423 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.414119 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.428642 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-logs\") pod \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.432809 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data-custom\") pod \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.433020 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bbxw\" (UniqueName: \"kubernetes.io/projected/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-kube-api-access-5bbxw\") pod \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.429351 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-logs" (OuterVolumeSpecName: "logs") pod "9b5721ed-4849-42c5-90c8-5524bcd5d9e1" (UID: "9b5721ed-4849-42c5-90c8-5524bcd5d9e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.433743 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-combined-ca-bundle\") pod \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.434992 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data\") pod \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\" (UID: \"9b5721ed-4849-42c5-90c8-5524bcd5d9e1\") " Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.436337 4954 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-logs\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.442290 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9b5721ed-4849-42c5-90c8-5524bcd5d9e1" (UID: "9b5721ed-4849-42c5-90c8-5524bcd5d9e1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.443646 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-kube-api-access-5bbxw" (OuterVolumeSpecName: "kube-api-access-5bbxw") pod "9b5721ed-4849-42c5-90c8-5524bcd5d9e1" (UID: "9b5721ed-4849-42c5-90c8-5524bcd5d9e1"). InnerVolumeSpecName "kube-api-access-5bbxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.521791 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b5721ed-4849-42c5-90c8-5524bcd5d9e1" (UID: "9b5721ed-4849-42c5-90c8-5524bcd5d9e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.526674 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-rsthw"] Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.527031 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" podUID="d0fe7343-f624-4a85-a1c2-1c05e4551159" containerName="dnsmasq-dns" containerID="cri-o://61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d" gracePeriod=10 Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.548721 4954 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.548764 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bbxw\" (UniqueName: \"kubernetes.io/projected/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-kube-api-access-5bbxw\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.548787 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.628763 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data" (OuterVolumeSpecName: "config-data") pod "9b5721ed-4849-42c5-90c8-5524bcd5d9e1" (UID: "9b5721ed-4849-42c5-90c8-5524bcd5d9e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.658764 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b5721ed-4849-42c5-90c8-5524bcd5d9e1-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.694303 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a8078c26-0c70-430b-b3d8-6f56d236657c","Type":"ContainerStarted","Data":"cfb6bd17d4826d10cbfbc82c3e0b966d20c5d9bd26f31a20173d2eca80b64c4f"} Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.696237 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.759147 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.759120744 podStartE2EDuration="3.759120744s" podCreationTimestamp="2025-12-09 17:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:07.739406707 +0000 UTC m=+1884.127580527" watchObservedRunningTime="2025-12-09 17:28:07.759120744 +0000 UTC m=+1884.147294564" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.781340 4954 generic.go:334] "Generic (PLEG): container finished" podID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerID="1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea" exitCode=0 Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.782763 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f98894b5d-v8d8l" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.782911 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f98894b5d-v8d8l" event={"ID":"9b5721ed-4849-42c5-90c8-5524bcd5d9e1","Type":"ContainerDied","Data":"1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea"} Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.782943 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f98894b5d-v8d8l" event={"ID":"9b5721ed-4849-42c5-90c8-5524bcd5d9e1","Type":"ContainerDied","Data":"001a939c1e5d78928b10c3058cb99317ab202be26f727ffda3c9089a8da562f6"} Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.782961 4954 scope.go:117] "RemoveContainer" containerID="1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea" Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.892412 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.949828 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-f98894b5d-v8d8l"] Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.967850 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-f98894b5d-v8d8l"] Dec 09 17:28:07 crc kubenswrapper[4954]: I1209 17:28:07.974546 4954 scope.go:117] "RemoveContainer" containerID="f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.021819 4954 scope.go:117] "RemoveContainer" containerID="1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea" Dec 09 17:28:08 crc kubenswrapper[4954]: E1209 17:28:08.025887 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea\": container with ID starting with 1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea not found: ID does not exist" containerID="1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.025958 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea"} err="failed to get container status \"1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea\": rpc error: code = NotFound desc = could not find container \"1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea\": container with ID starting with 1917461d4f29185275daab94ba0b8ea40d213170b7b10e4b6892e04255b6deea not found: ID does not exist" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.025993 4954 scope.go:117] "RemoveContainer" containerID="f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38" Dec 09 17:28:08 crc kubenswrapper[4954]: E1209 17:28:08.027467 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38\": container with ID starting with f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38 not found: ID does not exist" containerID="f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.027497 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38"} err="failed to get container status \"f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38\": rpc error: code = NotFound desc = could not find container \"f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38\": container with ID starting with f7c80d24b209ad66b3905f59e0d15cf7d5a1242a4603ab5e69f013eb68685a38 not found: ID does not exist" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.145348 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" path="/var/lib/kubelet/pods/9b5721ed-4849-42c5-90c8-5524bcd5d9e1/volumes" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.687735 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.802743 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l8pm\" (UniqueName: \"kubernetes.io/projected/d0fe7343-f624-4a85-a1c2-1c05e4551159-kube-api-access-5l8pm\") pod \"d0fe7343-f624-4a85-a1c2-1c05e4551159\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.803291 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-svc\") pod \"d0fe7343-f624-4a85-a1c2-1c05e4551159\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.803746 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-nb\") pod \"d0fe7343-f624-4a85-a1c2-1c05e4551159\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.803922 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-swift-storage-0\") pod \"d0fe7343-f624-4a85-a1c2-1c05e4551159\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.804106 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-sb\") pod \"d0fe7343-f624-4a85-a1c2-1c05e4551159\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.804648 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-config\") pod \"d0fe7343-f624-4a85-a1c2-1c05e4551159\" (UID: \"d0fe7343-f624-4a85-a1c2-1c05e4551159\") " Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.811281 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2224489a-740d-4225-bd71-f21cca0586f3","Type":"ContainerStarted","Data":"fd7914a9b690ffc823f05fc57f229f98743865220d38d754b7faf072837376dc"} Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.812060 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.831082 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0fe7343-f624-4a85-a1c2-1c05e4551159-kube-api-access-5l8pm" (OuterVolumeSpecName: "kube-api-access-5l8pm") pod "d0fe7343-f624-4a85-a1c2-1c05e4551159" (UID: "d0fe7343-f624-4a85-a1c2-1c05e4551159"). InnerVolumeSpecName "kube-api-access-5l8pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.842396 4954 generic.go:334] "Generic (PLEG): container finished" podID="d0fe7343-f624-4a85-a1c2-1c05e4551159" containerID="61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d" exitCode=0 Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.842821 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.843644 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" event={"ID":"d0fe7343-f624-4a85-a1c2-1c05e4551159","Type":"ContainerDied","Data":"61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d"} Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.843773 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-rsthw" event={"ID":"d0fe7343-f624-4a85-a1c2-1c05e4551159","Type":"ContainerDied","Data":"08fc02cd975a0f7b0145fab466c76f6b33b79f9b56f77e48bdb44de07b7e0293"} Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.843865 4954 scope.go:117] "RemoveContainer" containerID="61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.858728 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" containerName="probe" containerID="cri-o://de8b7c88c529ea3160f4ec26701963ce9790d28d8c82fc745e15191475a66b2d" gracePeriod=30 Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.868141 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" containerName="cinder-scheduler" containerID="cri-o://4a0c6b84827b9126618fe1522e0df3091d401ee36921bbefbc385b4ffa902807" gracePeriod=30 Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.882429 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.196792039 podStartE2EDuration="8.882395308s" podCreationTimestamp="2025-12-09 17:28:00 +0000 UTC" firstStartedPulling="2025-12-09 17:28:01.98072347 +0000 UTC m=+1878.368897290" lastFinishedPulling="2025-12-09 17:28:07.666326739 +0000 UTC m=+1884.054500559" observedRunningTime="2025-12-09 17:28:08.848121285 +0000 UTC m=+1885.236295105" watchObservedRunningTime="2025-12-09 17:28:08.882395308 +0000 UTC m=+1885.270569128" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.891910 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-config" (OuterVolumeSpecName: "config") pod "d0fe7343-f624-4a85-a1c2-1c05e4551159" (UID: "d0fe7343-f624-4a85-a1c2-1c05e4551159"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.907585 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.907683 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l8pm\" (UniqueName: \"kubernetes.io/projected/d0fe7343-f624-4a85-a1c2-1c05e4551159-kube-api-access-5l8pm\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.908772 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d0fe7343-f624-4a85-a1c2-1c05e4551159" (UID: "d0fe7343-f624-4a85-a1c2-1c05e4551159"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.910843 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d0fe7343-f624-4a85-a1c2-1c05e4551159" (UID: "d0fe7343-f624-4a85-a1c2-1c05e4551159"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.916731 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d0fe7343-f624-4a85-a1c2-1c05e4551159" (UID: "d0fe7343-f624-4a85-a1c2-1c05e4551159"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.924214 4954 scope.go:117] "RemoveContainer" containerID="fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318" Dec 09 17:28:08 crc kubenswrapper[4954]: I1209 17:28:08.937301 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d0fe7343-f624-4a85-a1c2-1c05e4551159" (UID: "d0fe7343-f624-4a85-a1c2-1c05e4551159"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.010819 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.010869 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.010883 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.010894 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0fe7343-f624-4a85-a1c2-1c05e4551159-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.041401 4954 scope.go:117] "RemoveContainer" containerID="61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d" Dec 09 17:28:09 crc kubenswrapper[4954]: E1209 17:28:09.042082 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d\": container with ID starting with 61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d not found: ID does not exist" containerID="61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.042236 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d"} err="failed to get container status \"61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d\": rpc error: code = NotFound desc = could not find container \"61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d\": container with ID starting with 61110cf1bfca9428a443c44ec86bc873a199b3ac0d04931da9d685519f82c19d not found: ID does not exist" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.042391 4954 scope.go:117] "RemoveContainer" containerID="fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318" Dec 09 17:28:09 crc kubenswrapper[4954]: E1209 17:28:09.043152 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318\": container with ID starting with fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318 not found: ID does not exist" containerID="fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.043282 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318"} err="failed to get container status \"fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318\": rpc error: code = NotFound desc = could not find container \"fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318\": container with ID starting with fce98842da10359a719cf91dde45daacf9dbe9cd8273d16b1a0dbbcf985a8318 not found: ID does not exist" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.195949 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-rsthw"] Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.208338 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-rsthw"] Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.218746 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lv87c"] Dec 09 17:28:09 crc kubenswrapper[4954]: E1209 17:28:09.219311 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerName="barbican-api" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.219334 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerName="barbican-api" Dec 09 17:28:09 crc kubenswrapper[4954]: E1209 17:28:09.219377 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fe7343-f624-4a85-a1c2-1c05e4551159" containerName="init" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.219384 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fe7343-f624-4a85-a1c2-1c05e4551159" containerName="init" Dec 09 17:28:09 crc kubenswrapper[4954]: E1209 17:28:09.219397 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerName="barbican-api-log" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.219404 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerName="barbican-api-log" Dec 09 17:28:09 crc kubenswrapper[4954]: E1209 17:28:09.219429 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fe7343-f624-4a85-a1c2-1c05e4551159" containerName="dnsmasq-dns" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.219435 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fe7343-f624-4a85-a1c2-1c05e4551159" containerName="dnsmasq-dns" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.219678 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerName="barbican-api" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.219697 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b5721ed-4849-42c5-90c8-5524bcd5d9e1" containerName="barbican-api-log" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.219737 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0fe7343-f624-4a85-a1c2-1c05e4551159" containerName="dnsmasq-dns" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.221463 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.230018 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lv87c"] Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.317004 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnm5b\" (UniqueName: \"kubernetes.io/projected/b512ca68-7a16-4948-91e4-76ced0037447-kube-api-access-qnm5b\") pod \"community-operators-lv87c\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.317073 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-utilities\") pod \"community-operators-lv87c\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.317098 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-catalog-content\") pod \"community-operators-lv87c\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.419482 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnm5b\" (UniqueName: \"kubernetes.io/projected/b512ca68-7a16-4948-91e4-76ced0037447-kube-api-access-qnm5b\") pod \"community-operators-lv87c\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.419583 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-utilities\") pod \"community-operators-lv87c\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.419633 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-catalog-content\") pod \"community-operators-lv87c\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.420277 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-utilities\") pod \"community-operators-lv87c\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.420301 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-catalog-content\") pod \"community-operators-lv87c\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.439506 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnm5b\" (UniqueName: \"kubernetes.io/projected/b512ca68-7a16-4948-91e4-76ced0037447-kube-api-access-qnm5b\") pod \"community-operators-lv87c\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.561039 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.925785 4954 generic.go:334] "Generic (PLEG): container finished" podID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" containerID="de8b7c88c529ea3160f4ec26701963ce9790d28d8c82fc745e15191475a66b2d" exitCode=0 Dec 09 17:28:09 crc kubenswrapper[4954]: I1209 17:28:09.927843 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6c7b7c8-3be0-4271-8ecd-f5db01209de7","Type":"ContainerDied","Data":"de8b7c88c529ea3160f4ec26701963ce9790d28d8c82fc745e15191475a66b2d"} Dec 09 17:28:10 crc kubenswrapper[4954]: I1209 17:28:10.134885 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0fe7343-f624-4a85-a1c2-1c05e4551159" path="/var/lib/kubelet/pods/d0fe7343-f624-4a85-a1c2-1c05e4551159/volumes" Dec 09 17:28:10 crc kubenswrapper[4954]: I1209 17:28:10.239464 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lv87c"] Dec 09 17:28:10 crc kubenswrapper[4954]: I1209 17:28:10.983282 4954 generic.go:334] "Generic (PLEG): container finished" podID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" containerID="4a0c6b84827b9126618fe1522e0df3091d401ee36921bbefbc385b4ffa902807" exitCode=0 Dec 09 17:28:10 crc kubenswrapper[4954]: I1209 17:28:10.983372 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6c7b7c8-3be0-4271-8ecd-f5db01209de7","Type":"ContainerDied","Data":"4a0c6b84827b9126618fe1522e0df3091d401ee36921bbefbc385b4ffa902807"} Dec 09 17:28:10 crc kubenswrapper[4954]: I1209 17:28:10.987183 4954 generic.go:334] "Generic (PLEG): container finished" podID="b512ca68-7a16-4948-91e4-76ced0037447" containerID="f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890" exitCode=0 Dec 09 17:28:10 crc kubenswrapper[4954]: I1209 17:28:10.987255 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv87c" event={"ID":"b512ca68-7a16-4948-91e4-76ced0037447","Type":"ContainerDied","Data":"f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890"} Dec 09 17:28:10 crc kubenswrapper[4954]: I1209 17:28:10.987334 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv87c" event={"ID":"b512ca68-7a16-4948-91e4-76ced0037447","Type":"ContainerStarted","Data":"e1d6fd0b79c8f382d0b51fab2a55db02fec033507a12d6f0c474aa7b08df624a"} Dec 09 17:28:10 crc kubenswrapper[4954]: I1209 17:28:10.991059 4954 generic.go:334] "Generic (PLEG): container finished" podID="284a949a-68ec-4470-a956-6e0a396675e0" containerID="58892b864944f9d101cc233d21c8b6e6f2e20e90c186a670c72216c198b1c92e" exitCode=0 Dec 09 17:28:10 crc kubenswrapper[4954]: I1209 17:28:10.991107 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76bd5dc95b-mwznb" event={"ID":"284a949a-68ec-4470-a956-6e0a396675e0","Type":"ContainerDied","Data":"58892b864944f9d101cc233d21c8b6e6f2e20e90c186a670c72216c198b1c92e"} Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.121065 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:28:11 crc kubenswrapper[4954]: E1209 17:28:11.121425 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.198167 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.221277 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twk69\" (UniqueName: \"kubernetes.io/projected/284a949a-68ec-4470-a956-6e0a396675e0-kube-api-access-twk69\") pod \"284a949a-68ec-4470-a956-6e0a396675e0\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.221349 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-config\") pod \"284a949a-68ec-4470-a956-6e0a396675e0\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.221411 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-ovndb-tls-certs\") pod \"284a949a-68ec-4470-a956-6e0a396675e0\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.221433 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-httpd-config\") pod \"284a949a-68ec-4470-a956-6e0a396675e0\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.221468 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-combined-ca-bundle\") pod \"284a949a-68ec-4470-a956-6e0a396675e0\" (UID: \"284a949a-68ec-4470-a956-6e0a396675e0\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.248076 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/284a949a-68ec-4470-a956-6e0a396675e0-kube-api-access-twk69" (OuterVolumeSpecName: "kube-api-access-twk69") pod "284a949a-68ec-4470-a956-6e0a396675e0" (UID: "284a949a-68ec-4470-a956-6e0a396675e0"). InnerVolumeSpecName "kube-api-access-twk69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.249105 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "284a949a-68ec-4470-a956-6e0a396675e0" (UID: "284a949a-68ec-4470-a956-6e0a396675e0"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.290179 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-config" (OuterVolumeSpecName: "config") pod "284a949a-68ec-4470-a956-6e0a396675e0" (UID: "284a949a-68ec-4470-a956-6e0a396675e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.324189 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twk69\" (UniqueName: \"kubernetes.io/projected/284a949a-68ec-4470-a956-6e0a396675e0-kube-api-access-twk69\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.324231 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.324244 4954 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.355993 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "284a949a-68ec-4470-a956-6e0a396675e0" (UID: "284a949a-68ec-4470-a956-6e0a396675e0"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.358275 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "284a949a-68ec-4470-a956-6e0a396675e0" (UID: "284a949a-68ec-4470-a956-6e0a396675e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.420237 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.425857 4954 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.425892 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/284a949a-68ec-4470-a956-6e0a396675e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.527736 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data-custom\") pod \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.528196 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-combined-ca-bundle\") pod \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.528251 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-scripts\") pod \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.528310 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data\") pod \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.528394 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-etc-machine-id\") pod \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.528669 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2wl4\" (UniqueName: \"kubernetes.io/projected/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-kube-api-access-n2wl4\") pod \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\" (UID: \"e6c7b7c8-3be0-4271-8ecd-f5db01209de7\") " Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.530490 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e6c7b7c8-3be0-4271-8ecd-f5db01209de7" (UID: "e6c7b7c8-3be0-4271-8ecd-f5db01209de7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.533661 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-kube-api-access-n2wl4" (OuterVolumeSpecName: "kube-api-access-n2wl4") pod "e6c7b7c8-3be0-4271-8ecd-f5db01209de7" (UID: "e6c7b7c8-3be0-4271-8ecd-f5db01209de7"). InnerVolumeSpecName "kube-api-access-n2wl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.533692 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-scripts" (OuterVolumeSpecName: "scripts") pod "e6c7b7c8-3be0-4271-8ecd-f5db01209de7" (UID: "e6c7b7c8-3be0-4271-8ecd-f5db01209de7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.538816 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e6c7b7c8-3be0-4271-8ecd-f5db01209de7" (UID: "e6c7b7c8-3be0-4271-8ecd-f5db01209de7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.594193 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6c7b7c8-3be0-4271-8ecd-f5db01209de7" (UID: "e6c7b7c8-3be0-4271-8ecd-f5db01209de7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.631722 4954 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.631766 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.631780 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.631791 4954 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.631802 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2wl4\" (UniqueName: \"kubernetes.io/projected/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-kube-api-access-n2wl4\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.642548 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data" (OuterVolumeSpecName: "config-data") pod "e6c7b7c8-3be0-4271-8ecd-f5db01209de7" (UID: "e6c7b7c8-3be0-4271-8ecd-f5db01209de7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:11 crc kubenswrapper[4954]: I1209 17:28:11.733864 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6c7b7c8-3be0-4271-8ecd-f5db01209de7-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.008898 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.008900 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6c7b7c8-3be0-4271-8ecd-f5db01209de7","Type":"ContainerDied","Data":"e50c5d15ad1b17cca20383c9d5ff2d88a9495cf4b52296ff93eef890289b6225"} Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.009457 4954 scope.go:117] "RemoveContainer" containerID="de8b7c88c529ea3160f4ec26701963ce9790d28d8c82fc745e15191475a66b2d" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.035722 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76bd5dc95b-mwznb" event={"ID":"284a949a-68ec-4470-a956-6e0a396675e0","Type":"ContainerDied","Data":"034e1ffc00381cea4cb662dba60369455c88d5efd04e71909d2f3c1fd0806f78"} Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.035813 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76bd5dc95b-mwznb" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.045060 4954 scope.go:117] "RemoveContainer" containerID="4a0c6b84827b9126618fe1522e0df3091d401ee36921bbefbc385b4ffa902807" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.071421 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.084103 4954 scope.go:117] "RemoveContainer" containerID="67698af236753d0e979ade240e6b3c96329e70ec61deb85d7206df9b666141d6" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.154207 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.154666 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-76bd5dc95b-mwznb"] Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.154688 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-76bd5dc95b-mwznb"] Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.170736 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 17:28:12 crc kubenswrapper[4954]: E1209 17:28:12.171488 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284a949a-68ec-4470-a956-6e0a396675e0" containerName="neutron-httpd" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.171510 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="284a949a-68ec-4470-a956-6e0a396675e0" containerName="neutron-httpd" Dec 09 17:28:12 crc kubenswrapper[4954]: E1209 17:28:12.171551 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" containerName="probe" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.171561 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" containerName="probe" Dec 09 17:28:12 crc kubenswrapper[4954]: E1209 17:28:12.171584 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284a949a-68ec-4470-a956-6e0a396675e0" containerName="neutron-api" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.171614 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="284a949a-68ec-4470-a956-6e0a396675e0" containerName="neutron-api" Dec 09 17:28:12 crc kubenswrapper[4954]: E1209 17:28:12.171649 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" containerName="cinder-scheduler" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.171655 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" containerName="cinder-scheduler" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.171953 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="284a949a-68ec-4470-a956-6e0a396675e0" containerName="neutron-httpd" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.171973 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" containerName="probe" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.172005 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" containerName="cinder-scheduler" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.172025 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="284a949a-68ec-4470-a956-6e0a396675e0" containerName="neutron-api" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.172773 4954 scope.go:117] "RemoveContainer" containerID="58892b864944f9d101cc233d21c8b6e6f2e20e90c186a670c72216c198b1c92e" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.173706 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.176879 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.185137 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.253839 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.254095 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.254200 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.254508 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzq9w\" (UniqueName: \"kubernetes.io/projected/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-kube-api-access-vzq9w\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.254542 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-scripts\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.254661 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-config-data\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.355643 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzq9w\" (UniqueName: \"kubernetes.io/projected/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-kube-api-access-vzq9w\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.355695 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-scripts\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.355739 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-config-data\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.355803 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.355839 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.355868 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.355994 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.365215 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-config-data\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.365658 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-scripts\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.383221 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.386238 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.388608 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzq9w\" (UniqueName: \"kubernetes.io/projected/74bc3bd9-6b4e-435c-b82e-3b1378962d6a-kube-api-access-vzq9w\") pod \"cinder-scheduler-0\" (UID: \"74bc3bd9-6b4e-435c-b82e-3b1378962d6a\") " pod="openstack/cinder-scheduler-0" Dec 09 17:28:12 crc kubenswrapper[4954]: I1209 17:28:12.516535 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 17:28:13 crc kubenswrapper[4954]: I1209 17:28:13.050140 4954 generic.go:334] "Generic (PLEG): container finished" podID="b512ca68-7a16-4948-91e4-76ced0037447" containerID="f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4" exitCode=0 Dec 09 17:28:13 crc kubenswrapper[4954]: I1209 17:28:13.050249 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv87c" event={"ID":"b512ca68-7a16-4948-91e4-76ced0037447","Type":"ContainerDied","Data":"f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4"} Dec 09 17:28:13 crc kubenswrapper[4954]: W1209 17:28:13.087327 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74bc3bd9_6b4e_435c_b82e_3b1378962d6a.slice/crio-ffb9d4301c5a8f565b04b7bcbb1e65e28df82ef8df7304f0d020b01484f3bcbf WatchSource:0}: Error finding container ffb9d4301c5a8f565b04b7bcbb1e65e28df82ef8df7304f0d020b01484f3bcbf: Status 404 returned error can't find the container with id ffb9d4301c5a8f565b04b7bcbb1e65e28df82ef8df7304f0d020b01484f3bcbf Dec 09 17:28:13 crc kubenswrapper[4954]: I1209 17:28:13.095727 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 17:28:14 crc kubenswrapper[4954]: I1209 17:28:14.104839 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv87c" event={"ID":"b512ca68-7a16-4948-91e4-76ced0037447","Type":"ContainerStarted","Data":"54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad"} Dec 09 17:28:14 crc kubenswrapper[4954]: I1209 17:28:14.145138 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lv87c" podStartSLOduration=2.533500702 podStartE2EDuration="5.145104449s" podCreationTimestamp="2025-12-09 17:28:09 +0000 UTC" firstStartedPulling="2025-12-09 17:28:10.993337722 +0000 UTC m=+1887.381511542" lastFinishedPulling="2025-12-09 17:28:13.604941469 +0000 UTC m=+1889.993115289" observedRunningTime="2025-12-09 17:28:14.126831367 +0000 UTC m=+1890.515005187" watchObservedRunningTime="2025-12-09 17:28:14.145104449 +0000 UTC m=+1890.533278269" Dec 09 17:28:14 crc kubenswrapper[4954]: I1209 17:28:14.159549 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="284a949a-68ec-4470-a956-6e0a396675e0" path="/var/lib/kubelet/pods/284a949a-68ec-4470-a956-6e0a396675e0/volumes" Dec 09 17:28:14 crc kubenswrapper[4954]: I1209 17:28:14.160425 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6c7b7c8-3be0-4271-8ecd-f5db01209de7" path="/var/lib/kubelet/pods/e6c7b7c8-3be0-4271-8ecd-f5db01209de7/volumes" Dec 09 17:28:14 crc kubenswrapper[4954]: I1209 17:28:14.162398 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"74bc3bd9-6b4e-435c-b82e-3b1378962d6a","Type":"ContainerStarted","Data":"a99f1c8d0f0638e2ba1751f3a8a5e833f7781756d0760a71e09321edb7e982ae"} Dec 09 17:28:14 crc kubenswrapper[4954]: I1209 17:28:14.162443 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"74bc3bd9-6b4e-435c-b82e-3b1378962d6a","Type":"ContainerStarted","Data":"ffb9d4301c5a8f565b04b7bcbb1e65e28df82ef8df7304f0d020b01484f3bcbf"} Dec 09 17:28:15 crc kubenswrapper[4954]: I1209 17:28:15.136702 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"74bc3bd9-6b4e-435c-b82e-3b1378962d6a","Type":"ContainerStarted","Data":"36953d6ad12155293678ffdccf633e7c5701ebbeb9f9a27554deb528bf428ab7"} Dec 09 17:28:15 crc kubenswrapper[4954]: I1209 17:28:15.170180 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.170158249 podStartE2EDuration="3.170158249s" podCreationTimestamp="2025-12-09 17:28:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:15.16509158 +0000 UTC m=+1891.553265400" watchObservedRunningTime="2025-12-09 17:28:15.170158249 +0000 UTC m=+1891.558332069" Dec 09 17:28:16 crc kubenswrapper[4954]: I1209 17:28:16.032013 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:28:16 crc kubenswrapper[4954]: I1209 17:28:16.102507 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-57bb47ccd8-dn5h5" Dec 09 17:28:17 crc kubenswrapper[4954]: I1209 17:28:17.516690 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 09 17:28:18 crc kubenswrapper[4954]: I1209 17:28:18.194053 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-64ffb74c6f-59w54" Dec 09 17:28:18 crc kubenswrapper[4954]: I1209 17:28:18.940431 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 09 17:28:19 crc kubenswrapper[4954]: I1209 17:28:19.561970 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:19 crc kubenswrapper[4954]: I1209 17:28:19.563045 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:19 crc kubenswrapper[4954]: I1209 17:28:19.613492 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:20 crc kubenswrapper[4954]: I1209 17:28:20.281497 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:21 crc kubenswrapper[4954]: I1209 17:28:21.942232 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 09 17:28:21 crc kubenswrapper[4954]: I1209 17:28:21.944153 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 09 17:28:21 crc kubenswrapper[4954]: I1209 17:28:21.947277 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 09 17:28:21 crc kubenswrapper[4954]: I1209 17:28:21.947282 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-gwrm6" Dec 09 17:28:21 crc kubenswrapper[4954]: I1209 17:28:21.947938 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 09 17:28:21 crc kubenswrapper[4954]: I1209 17:28:21.958151 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.027998 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-openstack-config-secret\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.028402 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj4tx\" (UniqueName: \"kubernetes.io/projected/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-kube-api-access-hj4tx\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.028498 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.028763 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-openstack-config\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.131527 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj4tx\" (UniqueName: \"kubernetes.io/projected/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-kube-api-access-hj4tx\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.131575 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.131636 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-openstack-config\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.131732 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-openstack-config-secret\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.133001 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-openstack-config\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.138839 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-openstack-config-secret\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.139860 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.155240 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj4tx\" (UniqueName: \"kubernetes.io/projected/6f8fc75c-fdd0-4931-9e61-02bccaee6b6d-kube-api-access-hj4tx\") pod \"openstackclient\" (UID: \"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d\") " pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.268749 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.808359 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 09 17:28:22 crc kubenswrapper[4954]: I1209 17:28:22.846733 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 09 17:28:23 crc kubenswrapper[4954]: I1209 17:28:23.120653 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:28:23 crc kubenswrapper[4954]: E1209 17:28:23.121179 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:28:23 crc kubenswrapper[4954]: I1209 17:28:23.253026 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d","Type":"ContainerStarted","Data":"78ef39217635e289303997463f6657315b221fa8e153ef558749e870b297ad9c"} Dec 09 17:28:24 crc kubenswrapper[4954]: I1209 17:28:24.384307 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lv87c"] Dec 09 17:28:24 crc kubenswrapper[4954]: I1209 17:28:24.384582 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lv87c" podUID="b512ca68-7a16-4948-91e4-76ced0037447" containerName="registry-server" containerID="cri-o://54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad" gracePeriod=2 Dec 09 17:28:24 crc kubenswrapper[4954]: I1209 17:28:24.969499 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.111055 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-catalog-content\") pod \"b512ca68-7a16-4948-91e4-76ced0037447\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.111160 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-utilities\") pod \"b512ca68-7a16-4948-91e4-76ced0037447\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.111478 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnm5b\" (UniqueName: \"kubernetes.io/projected/b512ca68-7a16-4948-91e4-76ced0037447-kube-api-access-qnm5b\") pod \"b512ca68-7a16-4948-91e4-76ced0037447\" (UID: \"b512ca68-7a16-4948-91e4-76ced0037447\") " Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.112397 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-utilities" (OuterVolumeSpecName: "utilities") pod "b512ca68-7a16-4948-91e4-76ced0037447" (UID: "b512ca68-7a16-4948-91e4-76ced0037447"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.118721 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b512ca68-7a16-4948-91e4-76ced0037447-kube-api-access-qnm5b" (OuterVolumeSpecName: "kube-api-access-qnm5b") pod "b512ca68-7a16-4948-91e4-76ced0037447" (UID: "b512ca68-7a16-4948-91e4-76ced0037447"). InnerVolumeSpecName "kube-api-access-qnm5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.185505 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b512ca68-7a16-4948-91e4-76ced0037447" (UID: "b512ca68-7a16-4948-91e4-76ced0037447"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.215446 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.215530 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b512ca68-7a16-4948-91e4-76ced0037447-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.215608 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnm5b\" (UniqueName: \"kubernetes.io/projected/b512ca68-7a16-4948-91e4-76ced0037447-kube-api-access-qnm5b\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.286480 4954 generic.go:334] "Generic (PLEG): container finished" podID="b512ca68-7a16-4948-91e4-76ced0037447" containerID="54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad" exitCode=0 Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.286538 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv87c" event={"ID":"b512ca68-7a16-4948-91e4-76ced0037447","Type":"ContainerDied","Data":"54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad"} Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.286550 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv87c" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.286579 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv87c" event={"ID":"b512ca68-7a16-4948-91e4-76ced0037447","Type":"ContainerDied","Data":"e1d6fd0b79c8f382d0b51fab2a55db02fec033507a12d6f0c474aa7b08df624a"} Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.286621 4954 scope.go:117] "RemoveContainer" containerID="54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.343787 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lv87c"] Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.353093 4954 scope.go:117] "RemoveContainer" containerID="f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.359908 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lv87c"] Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.395418 4954 scope.go:117] "RemoveContainer" containerID="f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.443526 4954 scope.go:117] "RemoveContainer" containerID="54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad" Dec 09 17:28:25 crc kubenswrapper[4954]: E1209 17:28:25.443905 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad\": container with ID starting with 54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad not found: ID does not exist" containerID="54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.443941 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad"} err="failed to get container status \"54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad\": rpc error: code = NotFound desc = could not find container \"54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad\": container with ID starting with 54ef71af7bd9bed1eb9393e077956189c05a40a4ec11ef5448db3268fc82dcad not found: ID does not exist" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.443970 4954 scope.go:117] "RemoveContainer" containerID="f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4" Dec 09 17:28:25 crc kubenswrapper[4954]: E1209 17:28:25.444249 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4\": container with ID starting with f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4 not found: ID does not exist" containerID="f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.444286 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4"} err="failed to get container status \"f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4\": rpc error: code = NotFound desc = could not find container \"f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4\": container with ID starting with f28c8646842ccb570e159811302c72258d202ed989ac7f9f6428cfa28d164db4 not found: ID does not exist" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.444306 4954 scope.go:117] "RemoveContainer" containerID="f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890" Dec 09 17:28:25 crc kubenswrapper[4954]: E1209 17:28:25.444756 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890\": container with ID starting with f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890 not found: ID does not exist" containerID="f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890" Dec 09 17:28:25 crc kubenswrapper[4954]: I1209 17:28:25.444785 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890"} err="failed to get container status \"f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890\": rpc error: code = NotFound desc = could not find container \"f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890\": container with ID starting with f9f3071dd10abd756f302060f78617fe90629c88e37cf453d8d7a3afc641b890 not found: ID does not exist" Dec 09 17:28:26 crc kubenswrapper[4954]: I1209 17:28:26.136275 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b512ca68-7a16-4948-91e4-76ced0037447" path="/var/lib/kubelet/pods/b512ca68-7a16-4948-91e4-76ced0037447/volumes" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.094643 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-547f8767c7-hv854"] Dec 09 17:28:27 crc kubenswrapper[4954]: E1209 17:28:27.095802 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b512ca68-7a16-4948-91e4-76ced0037447" containerName="registry-server" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.095829 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b512ca68-7a16-4948-91e4-76ced0037447" containerName="registry-server" Dec 09 17:28:27 crc kubenswrapper[4954]: E1209 17:28:27.095859 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b512ca68-7a16-4948-91e4-76ced0037447" containerName="extract-utilities" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.095870 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b512ca68-7a16-4948-91e4-76ced0037447" containerName="extract-utilities" Dec 09 17:28:27 crc kubenswrapper[4954]: E1209 17:28:27.095897 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b512ca68-7a16-4948-91e4-76ced0037447" containerName="extract-content" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.095912 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b512ca68-7a16-4948-91e4-76ced0037447" containerName="extract-content" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.096172 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="b512ca68-7a16-4948-91e4-76ced0037447" containerName="registry-server" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.098037 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.100720 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.100843 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.100968 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.122419 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-547f8767c7-hv854"] Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.178687 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-etc-swift\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.178792 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-public-tls-certs\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.178816 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-log-httpd\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.178872 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-config-data\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.178892 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp5vk\" (UniqueName: \"kubernetes.io/projected/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-kube-api-access-kp5vk\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.178914 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-internal-tls-certs\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.178977 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-run-httpd\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.179072 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-combined-ca-bundle\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.281373 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-etc-swift\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.281479 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-public-tls-certs\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.281506 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-log-httpd\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.281565 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-config-data\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.281606 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp5vk\" (UniqueName: \"kubernetes.io/projected/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-kube-api-access-kp5vk\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.281646 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-internal-tls-certs\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.281708 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-run-httpd\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.281781 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-combined-ca-bundle\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.282741 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-log-httpd\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.282811 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-run-httpd\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.290043 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-internal-tls-certs\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.290437 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-public-tls-certs\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.291049 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-etc-swift\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.291888 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-config-data\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.293945 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-combined-ca-bundle\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.302778 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp5vk\" (UniqueName: \"kubernetes.io/projected/a6635c8e-d9a5-4034-ab5c-7fe96dc10c10-kube-api-access-kp5vk\") pod \"swift-proxy-547f8767c7-hv854\" (UID: \"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10\") " pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:27 crc kubenswrapper[4954]: I1209 17:28:27.479186 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:28 crc kubenswrapper[4954]: I1209 17:28:28.534319 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:28 crc kubenswrapper[4954]: I1209 17:28:28.535060 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="ceilometer-central-agent" containerID="cri-o://eec4ad66049242ba5deee324cf61411d2975c4d953a0269250822af455084537" gracePeriod=30 Dec 09 17:28:28 crc kubenswrapper[4954]: I1209 17:28:28.535249 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="proxy-httpd" containerID="cri-o://fd7914a9b690ffc823f05fc57f229f98743865220d38d754b7faf072837376dc" gracePeriod=30 Dec 09 17:28:28 crc kubenswrapper[4954]: I1209 17:28:28.535302 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="sg-core" containerID="cri-o://3f99c83fa80b3b9f05baf5a6d89cf0b52b220d2d4cecfbef8f7c1cbbe2e6eab0" gracePeriod=30 Dec 09 17:28:28 crc kubenswrapper[4954]: I1209 17:28:28.535342 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="ceilometer-notification-agent" containerID="cri-o://3b24806e1afa23c91c40dcd2b739d3c623595b8232defe75355958cb931fa427" gracePeriod=30 Dec 09 17:28:28 crc kubenswrapper[4954]: I1209 17:28:28.550779 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.195:3000/\": EOF" Dec 09 17:28:29 crc kubenswrapper[4954]: I1209 17:28:29.372704 4954 generic.go:334] "Generic (PLEG): container finished" podID="2224489a-740d-4225-bd71-f21cca0586f3" containerID="fd7914a9b690ffc823f05fc57f229f98743865220d38d754b7faf072837376dc" exitCode=0 Dec 09 17:28:29 crc kubenswrapper[4954]: I1209 17:28:29.372773 4954 generic.go:334] "Generic (PLEG): container finished" podID="2224489a-740d-4225-bd71-f21cca0586f3" containerID="3f99c83fa80b3b9f05baf5a6d89cf0b52b220d2d4cecfbef8f7c1cbbe2e6eab0" exitCode=2 Dec 09 17:28:29 crc kubenswrapper[4954]: I1209 17:28:29.372781 4954 generic.go:334] "Generic (PLEG): container finished" podID="2224489a-740d-4225-bd71-f21cca0586f3" containerID="eec4ad66049242ba5deee324cf61411d2975c4d953a0269250822af455084537" exitCode=0 Dec 09 17:28:29 crc kubenswrapper[4954]: I1209 17:28:29.372758 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2224489a-740d-4225-bd71-f21cca0586f3","Type":"ContainerDied","Data":"fd7914a9b690ffc823f05fc57f229f98743865220d38d754b7faf072837376dc"} Dec 09 17:28:29 crc kubenswrapper[4954]: I1209 17:28:29.372855 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2224489a-740d-4225-bd71-f21cca0586f3","Type":"ContainerDied","Data":"3f99c83fa80b3b9f05baf5a6d89cf0b52b220d2d4cecfbef8f7c1cbbe2e6eab0"} Dec 09 17:28:29 crc kubenswrapper[4954]: I1209 17:28:29.372871 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2224489a-740d-4225-bd71-f21cca0586f3","Type":"ContainerDied","Data":"eec4ad66049242ba5deee324cf61411d2975c4d953a0269250822af455084537"} Dec 09 17:28:31 crc kubenswrapper[4954]: I1209 17:28:31.134856 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.195:3000/\": dial tcp 10.217.0.195:3000: connect: connection refused" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.421174 4954 generic.go:334] "Generic (PLEG): container finished" podID="2224489a-740d-4225-bd71-f21cca0586f3" containerID="3b24806e1afa23c91c40dcd2b739d3c623595b8232defe75355958cb931fa427" exitCode=0 Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.421378 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2224489a-740d-4225-bd71-f21cca0586f3","Type":"ContainerDied","Data":"3b24806e1afa23c91c40dcd2b739d3c623595b8232defe75355958cb931fa427"} Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.619799 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.623168 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-547f8767c7-hv854"] Dec 09 17:28:33 crc kubenswrapper[4954]: W1209 17:28:33.625010 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6635c8e_d9a5_4034_ab5c_7fe96dc10c10.slice/crio-f2921cbc05b1c5a7e9e30f7eac064c66cdf7ff9d20c4ccbf3d9bd77d71887c60 WatchSource:0}: Error finding container f2921cbc05b1c5a7e9e30f7eac064c66cdf7ff9d20c4ccbf3d9bd77d71887c60: Status 404 returned error can't find the container with id f2921cbc05b1c5a7e9e30f7eac064c66cdf7ff9d20c4ccbf3d9bd77d71887c60 Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.784246 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-combined-ca-bundle\") pod \"2224489a-740d-4225-bd71-f21cca0586f3\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.784325 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhghg\" (UniqueName: \"kubernetes.io/projected/2224489a-740d-4225-bd71-f21cca0586f3-kube-api-access-nhghg\") pod \"2224489a-740d-4225-bd71-f21cca0586f3\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.784549 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-run-httpd\") pod \"2224489a-740d-4225-bd71-f21cca0586f3\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.784654 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-log-httpd\") pod \"2224489a-740d-4225-bd71-f21cca0586f3\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.784704 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-scripts\") pod \"2224489a-740d-4225-bd71-f21cca0586f3\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.784773 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-config-data\") pod \"2224489a-740d-4225-bd71-f21cca0586f3\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.784867 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-sg-core-conf-yaml\") pod \"2224489a-740d-4225-bd71-f21cca0586f3\" (UID: \"2224489a-740d-4225-bd71-f21cca0586f3\") " Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.785173 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2224489a-740d-4225-bd71-f21cca0586f3" (UID: "2224489a-740d-4225-bd71-f21cca0586f3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.785399 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2224489a-740d-4225-bd71-f21cca0586f3" (UID: "2224489a-740d-4225-bd71-f21cca0586f3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.786099 4954 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.786123 4954 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2224489a-740d-4225-bd71-f21cca0586f3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.789463 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2224489a-740d-4225-bd71-f21cca0586f3-kube-api-access-nhghg" (OuterVolumeSpecName: "kube-api-access-nhghg") pod "2224489a-740d-4225-bd71-f21cca0586f3" (UID: "2224489a-740d-4225-bd71-f21cca0586f3"). InnerVolumeSpecName "kube-api-access-nhghg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.789734 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-scripts" (OuterVolumeSpecName: "scripts") pod "2224489a-740d-4225-bd71-f21cca0586f3" (UID: "2224489a-740d-4225-bd71-f21cca0586f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.828466 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2224489a-740d-4225-bd71-f21cca0586f3" (UID: "2224489a-740d-4225-bd71-f21cca0586f3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.888607 4954 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.888643 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhghg\" (UniqueName: \"kubernetes.io/projected/2224489a-740d-4225-bd71-f21cca0586f3-kube-api-access-nhghg\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.888656 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.894780 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2224489a-740d-4225-bd71-f21cca0586f3" (UID: "2224489a-740d-4225-bd71-f21cca0586f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.924636 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-config-data" (OuterVolumeSpecName: "config-data") pod "2224489a-740d-4225-bd71-f21cca0586f3" (UID: "2224489a-740d-4225-bd71-f21cca0586f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.991199 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:33 crc kubenswrapper[4954]: I1209 17:28:33.991239 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2224489a-740d-4225-bd71-f21cca0586f3-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.436103 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-547f8767c7-hv854" event={"ID":"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10","Type":"ContainerStarted","Data":"b3f32f589aeac165423f24689b20cd3d5ac27bb24a359a88219f8018467c6c9e"} Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.436516 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-547f8767c7-hv854" event={"ID":"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10","Type":"ContainerStarted","Data":"f1bcc2ffb9e834aa3505a1429fd50b0076664039b4033424e59c67b6882af716"} Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.436534 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-547f8767c7-hv854" event={"ID":"a6635c8e-d9a5-4034-ab5c-7fe96dc10c10","Type":"ContainerStarted","Data":"f2921cbc05b1c5a7e9e30f7eac064c66cdf7ff9d20c4ccbf3d9bd77d71887c60"} Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.436858 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.439275 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6f8fc75c-fdd0-4931-9e61-02bccaee6b6d","Type":"ContainerStarted","Data":"3f93746ae86d0a40219c430855f5b18bf42eb438a971cd5b8cfbcebf4710d034"} Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.451380 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2224489a-740d-4225-bd71-f21cca0586f3","Type":"ContainerDied","Data":"b144a2f4766e0ec40f31400d3e224f3bb179dca3c5733f18d89e838578cb0c90"} Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.451450 4954 scope.go:117] "RemoveContainer" containerID="fd7914a9b690ffc823f05fc57f229f98743865220d38d754b7faf072837376dc" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.451720 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.509529 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-547f8767c7-hv854" podStartSLOduration=7.509506086 podStartE2EDuration="7.509506086s" podCreationTimestamp="2025-12-09 17:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:34.48118687 +0000 UTC m=+1910.869360710" watchObservedRunningTime="2025-12-09 17:28:34.509506086 +0000 UTC m=+1910.897679906" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.524272 4954 scope.go:117] "RemoveContainer" containerID="3f99c83fa80b3b9f05baf5a6d89cf0b52b220d2d4cecfbef8f7c1cbbe2e6eab0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.535745 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.569072 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.576297 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.308063714 podStartE2EDuration="13.576270406s" podCreationTimestamp="2025-12-09 17:28:21 +0000 UTC" firstStartedPulling="2025-12-09 17:28:22.824960286 +0000 UTC m=+1899.213134106" lastFinishedPulling="2025-12-09 17:28:33.093166978 +0000 UTC m=+1909.481340798" observedRunningTime="2025-12-09 17:28:34.530635377 +0000 UTC m=+1910.918809207" watchObservedRunningTime="2025-12-09 17:28:34.576270406 +0000 UTC m=+1910.964444226" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.583886 4954 scope.go:117] "RemoveContainer" containerID="3b24806e1afa23c91c40dcd2b739d3c623595b8232defe75355958cb931fa427" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.592873 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:34 crc kubenswrapper[4954]: E1209 17:28:34.593475 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="proxy-httpd" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.593496 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="proxy-httpd" Dec 09 17:28:34 crc kubenswrapper[4954]: E1209 17:28:34.593518 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="ceilometer-notification-agent" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.593525 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="ceilometer-notification-agent" Dec 09 17:28:34 crc kubenswrapper[4954]: E1209 17:28:34.593556 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="ceilometer-central-agent" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.593563 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="ceilometer-central-agent" Dec 09 17:28:34 crc kubenswrapper[4954]: E1209 17:28:34.593617 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="sg-core" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.593625 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="sg-core" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.593833 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="sg-core" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.593864 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="ceilometer-central-agent" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.593881 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="proxy-httpd" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.593898 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="2224489a-740d-4225-bd71-f21cca0586f3" containerName="ceilometer-notification-agent" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.597729 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.603224 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.604248 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.604334 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.616182 4954 scope.go:117] "RemoveContainer" containerID="eec4ad66049242ba5deee324cf61411d2975c4d953a0269250822af455084537" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.710084 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z7k9\" (UniqueName: \"kubernetes.io/projected/f40981fe-a91e-45ef-a39c-ef449c19c751-kube-api-access-8z7k9\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.710149 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-run-httpd\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.710222 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-scripts\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.710385 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-config-data\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.710424 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.710456 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-log-httpd\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.710487 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.812329 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z7k9\" (UniqueName: \"kubernetes.io/projected/f40981fe-a91e-45ef-a39c-ef449c19c751-kube-api-access-8z7k9\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.812407 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-run-httpd\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.812472 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-scripts\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.812573 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-config-data\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.812632 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.812665 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-log-httpd\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.812698 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.813039 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-run-httpd\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.813741 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-log-httpd\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.819822 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.820315 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.820474 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-scripts\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.821175 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-config-data\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.843516 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z7k9\" (UniqueName: \"kubernetes.io/projected/f40981fe-a91e-45ef-a39c-ef449c19c751-kube-api-access-8z7k9\") pod \"ceilometer-0\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.923758 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:28:34 crc kubenswrapper[4954]: I1209 17:28:34.986740 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:35 crc kubenswrapper[4954]: I1209 17:28:35.417475 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:28:35 crc kubenswrapper[4954]: W1209 17:28:35.425231 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf40981fe_a91e_45ef_a39c_ef449c19c751.slice/crio-c40b4d4ffab2164339fd9b39a0f189ae1c24e0ae7e498e26304d31503c1866f6 WatchSource:0}: Error finding container c40b4d4ffab2164339fd9b39a0f189ae1c24e0ae7e498e26304d31503c1866f6: Status 404 returned error can't find the container with id c40b4d4ffab2164339fd9b39a0f189ae1c24e0ae7e498e26304d31503c1866f6 Dec 09 17:28:35 crc kubenswrapper[4954]: I1209 17:28:35.616734 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40981fe-a91e-45ef-a39c-ef449c19c751","Type":"ContainerStarted","Data":"c40b4d4ffab2164339fd9b39a0f189ae1c24e0ae7e498e26304d31503c1866f6"} Dec 09 17:28:35 crc kubenswrapper[4954]: I1209 17:28:35.617482 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:36 crc kubenswrapper[4954]: I1209 17:28:36.135285 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2224489a-740d-4225-bd71-f21cca0586f3" path="/var/lib/kubelet/pods/2224489a-740d-4225-bd71-f21cca0586f3/volumes" Dec 09 17:28:36 crc kubenswrapper[4954]: I1209 17:28:36.628763 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40981fe-a91e-45ef-a39c-ef449c19c751","Type":"ContainerStarted","Data":"da7e20528413ebb59131126e1769f5a793003b872b669e19c22f86bc40727556"} Dec 09 17:28:37 crc kubenswrapper[4954]: I1209 17:28:37.644445 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40981fe-a91e-45ef-a39c-ef449c19c751","Type":"ContainerStarted","Data":"d39c771e3ee46639b0fd9a5c1a37b4b6219ca751e5bca6c01b8914c849f52c61"} Dec 09 17:28:38 crc kubenswrapper[4954]: I1209 17:28:38.122118 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:28:38 crc kubenswrapper[4954]: E1209 17:28:38.122463 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:28:39 crc kubenswrapper[4954]: I1209 17:28:39.685952 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40981fe-a91e-45ef-a39c-ef449c19c751","Type":"ContainerStarted","Data":"0a27558d4d698f6c77257057c2215c9ffcd635d5dec3a5dec26aef26c8a9d5ec"} Dec 09 17:28:39 crc kubenswrapper[4954]: I1209 17:28:39.942136 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-kdt2k"] Dec 09 17:28:39 crc kubenswrapper[4954]: I1209 17:28:39.944246 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kdt2k" Dec 09 17:28:39 crc kubenswrapper[4954]: I1209 17:28:39.964501 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-kdt2k"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.062058 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7b79757c4c-zrc6v"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.073159 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.073165 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stwx5\" (UniqueName: \"kubernetes.io/projected/10f8b023-726d-443f-b728-eaa574d45bab-kube-api-access-stwx5\") pod \"nova-api-db-create-kdt2k\" (UID: \"10f8b023-726d-443f-b728-eaa574d45bab\") " pod="openstack/nova-api-db-create-kdt2k" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.073967 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10f8b023-726d-443f-b728-eaa574d45bab-operator-scripts\") pod \"nova-api-db-create-kdt2k\" (UID: \"10f8b023-726d-443f-b728-eaa574d45bab\") " pod="openstack/nova-api-db-create-kdt2k" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.077340 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.077703 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.077917 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-xcmzr" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.120706 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-z88n9"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.123267 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z88n9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.178621 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msctm\" (UniqueName: \"kubernetes.io/projected/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-kube-api-access-msctm\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.178707 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stwx5\" (UniqueName: \"kubernetes.io/projected/10f8b023-726d-443f-b728-eaa574d45bab-kube-api-access-stwx5\") pod \"nova-api-db-create-kdt2k\" (UID: \"10f8b023-726d-443f-b728-eaa574d45bab\") " pod="openstack/nova-api-db-create-kdt2k" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.178751 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqqr5\" (UniqueName: \"kubernetes.io/projected/41a8d52b-549b-42d1-91bf-e7e4567affb3-kube-api-access-lqqr5\") pod \"nova-cell0-db-create-z88n9\" (UID: \"41a8d52b-549b-42d1-91bf-e7e4567affb3\") " pod="openstack/nova-cell0-db-create-z88n9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.178772 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-combined-ca-bundle\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.178815 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10f8b023-726d-443f-b728-eaa574d45bab-operator-scripts\") pod \"nova-api-db-create-kdt2k\" (UID: \"10f8b023-726d-443f-b728-eaa574d45bab\") " pod="openstack/nova-api-db-create-kdt2k" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.178883 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.178916 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data-custom\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.178941 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41a8d52b-549b-42d1-91bf-e7e4567affb3-operator-scripts\") pod \"nova-cell0-db-create-z88n9\" (UID: \"41a8d52b-549b-42d1-91bf-e7e4567affb3\") " pod="openstack/nova-cell0-db-create-z88n9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.180430 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10f8b023-726d-443f-b728-eaa574d45bab-operator-scripts\") pod \"nova-api-db-create-kdt2k\" (UID: \"10f8b023-726d-443f-b728-eaa574d45bab\") " pod="openstack/nova-api-db-create-kdt2k" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.202067 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-z88n9"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.205397 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7b79757c4c-zrc6v"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.230612 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stwx5\" (UniqueName: \"kubernetes.io/projected/10f8b023-726d-443f-b728-eaa574d45bab-kube-api-access-stwx5\") pod \"nova-api-db-create-kdt2k\" (UID: \"10f8b023-726d-443f-b728-eaa574d45bab\") " pod="openstack/nova-api-db-create-kdt2k" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.266158 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kdt2k" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.280697 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msctm\" (UniqueName: \"kubernetes.io/projected/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-kube-api-access-msctm\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.280791 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqqr5\" (UniqueName: \"kubernetes.io/projected/41a8d52b-549b-42d1-91bf-e7e4567affb3-kube-api-access-lqqr5\") pod \"nova-cell0-db-create-z88n9\" (UID: \"41a8d52b-549b-42d1-91bf-e7e4567affb3\") " pod="openstack/nova-cell0-db-create-z88n9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.280829 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-combined-ca-bundle\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.280936 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.280980 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data-custom\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.281009 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41a8d52b-549b-42d1-91bf-e7e4567affb3-operator-scripts\") pod \"nova-cell0-db-create-z88n9\" (UID: \"41a8d52b-549b-42d1-91bf-e7e4567affb3\") " pod="openstack/nova-cell0-db-create-z88n9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.284236 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41a8d52b-549b-42d1-91bf-e7e4567affb3-operator-scripts\") pod \"nova-cell0-db-create-z88n9\" (UID: \"41a8d52b-549b-42d1-91bf-e7e4567affb3\") " pod="openstack/nova-cell0-db-create-z88n9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.294466 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-c84cd"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.296405 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c84cd" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.305179 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-combined-ca-bundle\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.305397 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.311739 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data-custom\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.327160 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msctm\" (UniqueName: \"kubernetes.io/projected/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-kube-api-access-msctm\") pod \"heat-engine-7b79757c4c-zrc6v\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.347319 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqqr5\" (UniqueName: \"kubernetes.io/projected/41a8d52b-549b-42d1-91bf-e7e4567affb3-kube-api-access-lqqr5\") pod \"nova-cell0-db-create-z88n9\" (UID: \"41a8d52b-549b-42d1-91bf-e7e4567affb3\") " pod="openstack/nova-cell0-db-create-z88n9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.367257 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-c84cd"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.383503 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5vd2\" (UniqueName: \"kubernetes.io/projected/d7d295cb-5d01-4014-abbe-efc80d8259f8-kube-api-access-f5vd2\") pod \"nova-cell1-db-create-c84cd\" (UID: \"d7d295cb-5d01-4014-abbe-efc80d8259f8\") " pod="openstack/nova-cell1-db-create-c84cd" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.383698 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7d295cb-5d01-4014-abbe-efc80d8259f8-operator-scripts\") pod \"nova-cell1-db-create-c84cd\" (UID: \"d7d295cb-5d01-4014-abbe-efc80d8259f8\") " pod="openstack/nova-cell1-db-create-c84cd" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.395957 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z88n9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.403340 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-9zsk9"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.410924 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.448656 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-9zsk9"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.469208 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7c8bd6d4b7-mg8lb"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.470962 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.481669 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.481686 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-a4ac-account-create-update-qprph"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.483682 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a4ac-account-create-update-qprph" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.488453 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7d295cb-5d01-4014-abbe-efc80d8259f8-operator-scripts\") pod \"nova-cell1-db-create-c84cd\" (UID: \"d7d295cb-5d01-4014-abbe-efc80d8259f8\") " pod="openstack/nova-cell1-db-create-c84cd" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.488503 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.488543 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.488587 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.488623 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-config\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.488697 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9687x\" (UniqueName: \"kubernetes.io/projected/5b762821-719b-451b-9b19-4f1ae1b12994-kube-api-access-9687x\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.488732 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.488777 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5vd2\" (UniqueName: \"kubernetes.io/projected/d7d295cb-5d01-4014-abbe-efc80d8259f8-kube-api-access-f5vd2\") pod \"nova-cell1-db-create-c84cd\" (UID: \"d7d295cb-5d01-4014-abbe-efc80d8259f8\") " pod="openstack/nova-cell1-db-create-c84cd" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.489925 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7d295cb-5d01-4014-abbe-efc80d8259f8-operator-scripts\") pod \"nova-cell1-db-create-c84cd\" (UID: \"d7d295cb-5d01-4014-abbe-efc80d8259f8\") " pod="openstack/nova-cell1-db-create-c84cd" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.500227 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.508108 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a4ac-account-create-update-qprph"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.542030 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5vd2\" (UniqueName: \"kubernetes.io/projected/d7d295cb-5d01-4014-abbe-efc80d8259f8-kube-api-access-f5vd2\") pod \"nova-cell1-db-create-c84cd\" (UID: \"d7d295cb-5d01-4014-abbe-efc80d8259f8\") " pod="openstack/nova-cell1-db-create-c84cd" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.542975 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7c8bd6d4b7-mg8lb"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.588571 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5f58db749d-zq2l7"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.592462 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.592541 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.592621 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.592641 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-config\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.592722 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9687x\" (UniqueName: \"kubernetes.io/projected/5b762821-719b-451b-9b19-4f1ae1b12994-kube-api-access-9687x\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.592771 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.593494 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.594352 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-config\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.595139 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.595701 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.595730 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.596359 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.599585 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.600332 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.625940 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5f58db749d-zq2l7"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.626108 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9687x\" (UniqueName: \"kubernetes.io/projected/5b762821-719b-451b-9b19-4f1ae1b12994-kube-api-access-9687x\") pod \"dnsmasq-dns-7756b9d78c-9zsk9\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.647979 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-602a-account-create-update-892sk"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.649680 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-602a-account-create-update-892sk" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.652799 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.665961 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-602a-account-create-update-892sk"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.695500 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data-custom\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.695638 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/588cbebb-ba39-4145-9c5e-865f77ee92cb-operator-scripts\") pod \"nova-api-a4ac-account-create-update-qprph\" (UID: \"588cbebb-ba39-4145-9c5e-865f77ee92cb\") " pod="openstack/nova-api-a4ac-account-create-update-qprph" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.695663 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-combined-ca-bundle\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.695693 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data-custom\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.695746 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.695785 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.695809 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-combined-ca-bundle\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.695834 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9whpx\" (UniqueName: \"kubernetes.io/projected/91fe1cc4-243c-473f-a86a-2a961648027f-kube-api-access-9whpx\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.695868 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8gbt\" (UniqueName: \"kubernetes.io/projected/588cbebb-ba39-4145-9c5e-865f77ee92cb-kube-api-access-j8gbt\") pod \"nova-api-a4ac-account-create-update-qprph\" (UID: \"588cbebb-ba39-4145-9c5e-865f77ee92cb\") " pod="openstack/nova-api-a4ac-account-create-update-qprph" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.695939 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvz94\" (UniqueName: \"kubernetes.io/projected/8b8a6386-1d71-4193-8d87-303561f4b34a-kube-api-access-lvz94\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.732888 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c84cd" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.751234 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40981fe-a91e-45ef-a39c-ef449c19c751","Type":"ContainerStarted","Data":"109b477bde6f46241718782f183745f21cb35376ea49165b7bd443649b0f5566"} Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.751469 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="ceilometer-central-agent" containerID="cri-o://da7e20528413ebb59131126e1769f5a793003b872b669e19c22f86bc40727556" gracePeriod=30 Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.751564 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.751765 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="proxy-httpd" containerID="cri-o://109b477bde6f46241718782f183745f21cb35376ea49165b7bd443649b0f5566" gracePeriod=30 Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.751832 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="ceilometer-notification-agent" containerID="cri-o://d39c771e3ee46639b0fd9a5c1a37b4b6219ca751e5bca6c01b8914c849f52c61" gracePeriod=30 Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.752006 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="sg-core" containerID="cri-o://0a27558d4d698f6c77257057c2215c9ffcd635d5dec3a5dec26aef26c8a9d5ec" gracePeriod=30 Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.775532 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-bce8-account-create-update-lhxt6"] Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.792114 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.799730 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.808297 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.857842 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8gbt\" (UniqueName: \"kubernetes.io/projected/588cbebb-ba39-4145-9c5e-865f77ee92cb-kube-api-access-j8gbt\") pod \"nova-api-a4ac-account-create-update-qprph\" (UID: \"588cbebb-ba39-4145-9c5e-865f77ee92cb\") " pod="openstack/nova-api-a4ac-account-create-update-qprph" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.858337 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1683bbec-ff24-4d14-aed8-cfb49b328355-operator-scripts\") pod \"nova-cell0-602a-account-create-update-892sk\" (UID: \"1683bbec-ff24-4d14-aed8-cfb49b328355\") " pod="openstack/nova-cell0-602a-account-create-update-892sk" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.858417 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvz94\" (UniqueName: \"kubernetes.io/projected/8b8a6386-1d71-4193-8d87-303561f4b34a-kube-api-access-lvz94\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.858618 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data-custom\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.858742 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqb4l\" (UniqueName: \"kubernetes.io/projected/1683bbec-ff24-4d14-aed8-cfb49b328355-kube-api-access-vqb4l\") pod \"nova-cell0-602a-account-create-update-892sk\" (UID: \"1683bbec-ff24-4d14-aed8-cfb49b328355\") " pod="openstack/nova-cell0-602a-account-create-update-892sk" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.858876 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/588cbebb-ba39-4145-9c5e-865f77ee92cb-operator-scripts\") pod \"nova-api-a4ac-account-create-update-qprph\" (UID: \"588cbebb-ba39-4145-9c5e-865f77ee92cb\") " pod="openstack/nova-api-a4ac-account-create-update-qprph" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.858919 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-combined-ca-bundle\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.858963 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data-custom\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.859059 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.859106 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.859141 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-combined-ca-bundle\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.859165 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9whpx\" (UniqueName: \"kubernetes.io/projected/91fe1cc4-243c-473f-a86a-2a961648027f-kube-api-access-9whpx\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.860778 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/588cbebb-ba39-4145-9c5e-865f77ee92cb-operator-scripts\") pod \"nova-api-a4ac-account-create-update-qprph\" (UID: \"588cbebb-ba39-4145-9c5e-865f77ee92cb\") " pod="openstack/nova-api-a4ac-account-create-update-qprph" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.925281 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-combined-ca-bundle\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.929810 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-combined-ca-bundle\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.943275 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.945062 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9whpx\" (UniqueName: \"kubernetes.io/projected/91fe1cc4-243c-473f-a86a-2a961648027f-kube-api-access-9whpx\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.950511 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.989033 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvz94\" (UniqueName: \"kubernetes.io/projected/8b8a6386-1d71-4193-8d87-303561f4b34a-kube-api-access-lvz94\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.989698 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data-custom\") pod \"heat-api-5f58db749d-zq2l7\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:40 crc kubenswrapper[4954]: I1209 17:28:40.995871 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-bce8-account-create-update-lhxt6"] Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.007434 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data-custom\") pod \"heat-cfnapi-7c8bd6d4b7-mg8lb\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.008267 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1683bbec-ff24-4d14-aed8-cfb49b328355-operator-scripts\") pod \"nova-cell0-602a-account-create-update-892sk\" (UID: \"1683bbec-ff24-4d14-aed8-cfb49b328355\") " pod="openstack/nova-cell0-602a-account-create-update-892sk" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.008471 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjgc9\" (UniqueName: \"kubernetes.io/projected/f599c00a-fbf5-46cb-9304-f5554664eb3e-kube-api-access-tjgc9\") pod \"nova-cell1-bce8-account-create-update-lhxt6\" (UID: \"f599c00a-fbf5-46cb-9304-f5554664eb3e\") " pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.008727 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqb4l\" (UniqueName: \"kubernetes.io/projected/1683bbec-ff24-4d14-aed8-cfb49b328355-kube-api-access-vqb4l\") pod \"nova-cell0-602a-account-create-update-892sk\" (UID: \"1683bbec-ff24-4d14-aed8-cfb49b328355\") " pod="openstack/nova-cell0-602a-account-create-update-892sk" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.008915 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f599c00a-fbf5-46cb-9304-f5554664eb3e-operator-scripts\") pod \"nova-cell1-bce8-account-create-update-lhxt6\" (UID: \"f599c00a-fbf5-46cb-9304-f5554664eb3e\") " pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.010527 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1683bbec-ff24-4d14-aed8-cfb49b328355-operator-scripts\") pod \"nova-cell0-602a-account-create-update-892sk\" (UID: \"1683bbec-ff24-4d14-aed8-cfb49b328355\") " pod="openstack/nova-cell0-602a-account-create-update-892sk" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.032955 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8gbt\" (UniqueName: \"kubernetes.io/projected/588cbebb-ba39-4145-9c5e-865f77ee92cb-kube-api-access-j8gbt\") pod \"nova-api-a4ac-account-create-update-qprph\" (UID: \"588cbebb-ba39-4145-9c5e-865f77ee92cb\") " pod="openstack/nova-api-a4ac-account-create-update-qprph" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.047242 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqb4l\" (UniqueName: \"kubernetes.io/projected/1683bbec-ff24-4d14-aed8-cfb49b328355-kube-api-access-vqb4l\") pod \"nova-cell0-602a-account-create-update-892sk\" (UID: \"1683bbec-ff24-4d14-aed8-cfb49b328355\") " pod="openstack/nova-cell0-602a-account-create-update-892sk" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.090232 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.476946808 podStartE2EDuration="7.090185537s" podCreationTimestamp="2025-12-09 17:28:34 +0000 UTC" firstStartedPulling="2025-12-09 17:28:35.428104854 +0000 UTC m=+1911.816278664" lastFinishedPulling="2025-12-09 17:28:40.041343573 +0000 UTC m=+1916.429517393" observedRunningTime="2025-12-09 17:28:40.883330631 +0000 UTC m=+1917.271504461" watchObservedRunningTime="2025-12-09 17:28:41.090185537 +0000 UTC m=+1917.478359357" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.106979 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.112191 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f599c00a-fbf5-46cb-9304-f5554664eb3e-operator-scripts\") pod \"nova-cell1-bce8-account-create-update-lhxt6\" (UID: \"f599c00a-fbf5-46cb-9304-f5554664eb3e\") " pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.112330 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjgc9\" (UniqueName: \"kubernetes.io/projected/f599c00a-fbf5-46cb-9304-f5554664eb3e-kube-api-access-tjgc9\") pod \"nova-cell1-bce8-account-create-update-lhxt6\" (UID: \"f599c00a-fbf5-46cb-9304-f5554664eb3e\") " pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.113409 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f599c00a-fbf5-46cb-9304-f5554664eb3e-operator-scripts\") pod \"nova-cell1-bce8-account-create-update-lhxt6\" (UID: \"f599c00a-fbf5-46cb-9304-f5554664eb3e\") " pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.142964 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjgc9\" (UniqueName: \"kubernetes.io/projected/f599c00a-fbf5-46cb-9304-f5554664eb3e-kube-api-access-tjgc9\") pod \"nova-cell1-bce8-account-create-update-lhxt6\" (UID: \"f599c00a-fbf5-46cb-9304-f5554664eb3e\") " pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.143621 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a4ac-account-create-update-qprph" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.157607 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.221514 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.277072 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-602a-account-create-update-892sk" Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.379487 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-kdt2k"] Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.400043 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-z88n9"] Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.626297 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7b79757c4c-zrc6v"] Dec 09 17:28:41 crc kubenswrapper[4954]: W1209 17:28:41.659810 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfe2f3db_ef77_459d_8ee8_e0df346b5beb.slice/crio-78ca2073596ce3e4710dfa2bd77ca138fff6555838919904681917c5209bce6f WatchSource:0}: Error finding container 78ca2073596ce3e4710dfa2bd77ca138fff6555838919904681917c5209bce6f: Status 404 returned error can't find the container with id 78ca2073596ce3e4710dfa2bd77ca138fff6555838919904681917c5209bce6f Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.799295 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z88n9" event={"ID":"41a8d52b-549b-42d1-91bf-e7e4567affb3","Type":"ContainerStarted","Data":"c7a3ffb103fef31466f1074b75916ad6309e0a728625c5b16259c274971e1f8b"} Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.819555 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kdt2k" event={"ID":"10f8b023-726d-443f-b728-eaa574d45bab","Type":"ContainerStarted","Data":"00cb7d8efd62d1f867747365e8abb3fd8e1292d542bf3326e12dbc44c0b37f8f"} Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.830400 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-9zsk9"] Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.833889 4954 generic.go:334] "Generic (PLEG): container finished" podID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerID="0a27558d4d698f6c77257057c2215c9ffcd635d5dec3a5dec26aef26c8a9d5ec" exitCode=2 Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.833952 4954 generic.go:334] "Generic (PLEG): container finished" podID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerID="d39c771e3ee46639b0fd9a5c1a37b4b6219ca751e5bca6c01b8914c849f52c61" exitCode=0 Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.834075 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40981fe-a91e-45ef-a39c-ef449c19c751","Type":"ContainerDied","Data":"0a27558d4d698f6c77257057c2215c9ffcd635d5dec3a5dec26aef26c8a9d5ec"} Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.834127 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40981fe-a91e-45ef-a39c-ef449c19c751","Type":"ContainerDied","Data":"d39c771e3ee46639b0fd9a5c1a37b4b6219ca751e5bca6c01b8914c849f52c61"} Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.837882 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7b79757c4c-zrc6v" event={"ID":"cfe2f3db-ef77-459d-8ee8-e0df346b5beb","Type":"ContainerStarted","Data":"78ca2073596ce3e4710dfa2bd77ca138fff6555838919904681917c5209bce6f"} Dec 09 17:28:41 crc kubenswrapper[4954]: I1209 17:28:41.849662 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-c84cd"] Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.086533 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-bce8-account-create-update-lhxt6"] Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.115755 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7c8bd6d4b7-mg8lb"] Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.368502 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5f58db749d-zq2l7"] Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.382618 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a4ac-account-create-update-qprph"] Dec 09 17:28:42 crc kubenswrapper[4954]: W1209 17:28:42.386361 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod588cbebb_ba39_4145_9c5e_865f77ee92cb.slice/crio-79329071ad481ea57a0852de3ac42f2d3cc754b13044c9e48918aed5646b044e WatchSource:0}: Error finding container 79329071ad481ea57a0852de3ac42f2d3cc754b13044c9e48918aed5646b044e: Status 404 returned error can't find the container with id 79329071ad481ea57a0852de3ac42f2d3cc754b13044c9e48918aed5646b044e Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.491205 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.502107 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-547f8767c7-hv854" Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.553786 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-602a-account-create-update-892sk"] Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.861193 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" event={"ID":"91fe1cc4-243c-473f-a86a-2a961648027f","Type":"ContainerStarted","Data":"6ddfd65caf3219c2dfb444d2eb22bc682fc0c6b9ce693d51da113bdebd3768f4"} Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.874780 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-c84cd" event={"ID":"d7d295cb-5d01-4014-abbe-efc80d8259f8","Type":"ContainerStarted","Data":"019944e8bf6e1f3b40a622ec027ff5dd4864764b4b0a13313c3f50e00cb28fd5"} Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.874838 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-c84cd" event={"ID":"d7d295cb-5d01-4014-abbe-efc80d8259f8","Type":"ContainerStarted","Data":"f212cc5e7475c0facaee6067e00af723e6b1e513b72f6564b335bb215823add2"} Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.878310 4954 generic.go:334] "Generic (PLEG): container finished" podID="10f8b023-726d-443f-b728-eaa574d45bab" containerID="b612b367e033820f2f8473c0faf858404b5431b61e6e8f0a12445111edf8be64" exitCode=0 Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.878430 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kdt2k" event={"ID":"10f8b023-726d-443f-b728-eaa574d45bab","Type":"ContainerDied","Data":"b612b367e033820f2f8473c0faf858404b5431b61e6e8f0a12445111edf8be64"} Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.888798 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a4ac-account-create-update-qprph" event={"ID":"588cbebb-ba39-4145-9c5e-865f77ee92cb","Type":"ContainerStarted","Data":"79329071ad481ea57a0852de3ac42f2d3cc754b13044c9e48918aed5646b044e"} Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.919964 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7b79757c4c-zrc6v" event={"ID":"cfe2f3db-ef77-459d-8ee8-e0df346b5beb","Type":"ContainerStarted","Data":"63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e"} Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.921237 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.923856 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-c84cd" podStartSLOduration=2.92382977 podStartE2EDuration="2.92382977s" podCreationTimestamp="2025-12-09 17:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:42.905246549 +0000 UTC m=+1919.293420389" watchObservedRunningTime="2025-12-09 17:28:42.92382977 +0000 UTC m=+1919.312003590" Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.950544 4954 generic.go:334] "Generic (PLEG): container finished" podID="41a8d52b-549b-42d1-91bf-e7e4567affb3" containerID="3ef13b2db96f3a68ca5e0d44d2301257ab37079cdd3f1aaa08a84bb45e3733d8" exitCode=0 Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.951187 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z88n9" event={"ID":"41a8d52b-549b-42d1-91bf-e7e4567affb3","Type":"ContainerDied","Data":"3ef13b2db96f3a68ca5e0d44d2301257ab37079cdd3f1aaa08a84bb45e3733d8"} Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.952475 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-a4ac-account-create-update-qprph" podStartSLOduration=2.952451866 podStartE2EDuration="2.952451866s" podCreationTimestamp="2025-12-09 17:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:42.946453758 +0000 UTC m=+1919.334627578" watchObservedRunningTime="2025-12-09 17:28:42.952451866 +0000 UTC m=+1919.340625696" Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.970854 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-602a-account-create-update-892sk" event={"ID":"1683bbec-ff24-4d14-aed8-cfb49b328355","Type":"ContainerStarted","Data":"130e33fa5950599aa47163d6466aed92d8f1e7fbf5371e130c6791d4d2111ca0"} Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.987215 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" event={"ID":"f599c00a-fbf5-46cb-9304-f5554664eb3e","Type":"ContainerStarted","Data":"013ac2deeccccad2e28d9d3ccdba7f6af5648419987820c5543ba5893f5717c2"} Dec 09 17:28:42 crc kubenswrapper[4954]: I1209 17:28:42.987268 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" event={"ID":"f599c00a-fbf5-46cb-9304-f5554664eb3e","Type":"ContainerStarted","Data":"1108407dbd388d6e77dd6c15a950bf1a22e516ef8ac0e5a720539619be9e6243"} Dec 09 17:28:43 crc kubenswrapper[4954]: I1209 17:28:43.003706 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5f58db749d-zq2l7" event={"ID":"8b8a6386-1d71-4193-8d87-303561f4b34a","Type":"ContainerStarted","Data":"4311ac47319841ee55591d6a90ab452ac77a005742c5819b790ceb70942d054a"} Dec 09 17:28:43 crc kubenswrapper[4954]: I1209 17:28:43.066062 4954 generic.go:334] "Generic (PLEG): container finished" podID="5b762821-719b-451b-9b19-4f1ae1b12994" containerID="c77240abe49e549301e6ccb00a9526916075b35e9d421271452ee32df1edc332" exitCode=0 Dec 09 17:28:43 crc kubenswrapper[4954]: I1209 17:28:43.066356 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" event={"ID":"5b762821-719b-451b-9b19-4f1ae1b12994","Type":"ContainerDied","Data":"c77240abe49e549301e6ccb00a9526916075b35e9d421271452ee32df1edc332"} Dec 09 17:28:43 crc kubenswrapper[4954]: I1209 17:28:43.066409 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" event={"ID":"5b762821-719b-451b-9b19-4f1ae1b12994","Type":"ContainerStarted","Data":"f859ca87cc9a98a954010adfdd67557ab77376fdc7a4b8e8c1dd2ceb1d93ae7e"} Dec 09 17:28:43 crc kubenswrapper[4954]: I1209 17:28:43.147035 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7b79757c4c-zrc6v" podStartSLOduration=4.146994166 podStartE2EDuration="4.146994166s" podCreationTimestamp="2025-12-09 17:28:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:43.00335053 +0000 UTC m=+1919.391524350" watchObservedRunningTime="2025-12-09 17:28:43.146994166 +0000 UTC m=+1919.535167986" Dec 09 17:28:43 crc kubenswrapper[4954]: I1209 17:28:43.340568 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" podStartSLOduration=3.340534465 podStartE2EDuration="3.340534465s" podCreationTimestamp="2025-12-09 17:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:43.065847246 +0000 UTC m=+1919.454021066" watchObservedRunningTime="2025-12-09 17:28:43.340534465 +0000 UTC m=+1919.728708285" Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.080798 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" event={"ID":"5b762821-719b-451b-9b19-4f1ae1b12994","Type":"ContainerStarted","Data":"0bbe12244509bb8e93cb6731a09e52e88cadfef4588187c14193213cfbb07b4d"} Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.082255 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.098705 4954 generic.go:334] "Generic (PLEG): container finished" podID="d7d295cb-5d01-4014-abbe-efc80d8259f8" containerID="019944e8bf6e1f3b40a622ec027ff5dd4864764b4b0a13313c3f50e00cb28fd5" exitCode=0 Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.098873 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-c84cd" event={"ID":"d7d295cb-5d01-4014-abbe-efc80d8259f8","Type":"ContainerDied","Data":"019944e8bf6e1f3b40a622ec027ff5dd4864764b4b0a13313c3f50e00cb28fd5"} Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.112395 4954 generic.go:334] "Generic (PLEG): container finished" podID="1683bbec-ff24-4d14-aed8-cfb49b328355" containerID="dbde0662e80db3b450369d2715e456152bbd324324f5b0e040fae5521791b31c" exitCode=0 Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.112510 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-602a-account-create-update-892sk" event={"ID":"1683bbec-ff24-4d14-aed8-cfb49b328355","Type":"ContainerDied","Data":"dbde0662e80db3b450369d2715e456152bbd324324f5b0e040fae5521791b31c"} Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.131731 4954 generic.go:334] "Generic (PLEG): container finished" podID="588cbebb-ba39-4145-9c5e-865f77ee92cb" containerID="16c693de493d33217dab79116f369ab5b07d38b359ecc03353517560d4a21f6b" exitCode=0 Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.132626 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" podStartSLOduration=4.132605261 podStartE2EDuration="4.132605261s" podCreationTimestamp="2025-12-09 17:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:44.105654698 +0000 UTC m=+1920.493828538" watchObservedRunningTime="2025-12-09 17:28:44.132605261 +0000 UTC m=+1920.520779081" Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.139054 4954 generic.go:334] "Generic (PLEG): container finished" podID="f599c00a-fbf5-46cb-9304-f5554664eb3e" containerID="013ac2deeccccad2e28d9d3ccdba7f6af5648419987820c5543ba5893f5717c2" exitCode=0 Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.153346 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a4ac-account-create-update-qprph" event={"ID":"588cbebb-ba39-4145-9c5e-865f77ee92cb","Type":"ContainerDied","Data":"16c693de493d33217dab79116f369ab5b07d38b359ecc03353517560d4a21f6b"} Dec 09 17:28:44 crc kubenswrapper[4954]: I1209 17:28:44.153416 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" event={"ID":"f599c00a-fbf5-46cb-9304-f5554664eb3e","Type":"ContainerDied","Data":"013ac2deeccccad2e28d9d3ccdba7f6af5648419987820c5543ba5893f5717c2"} Dec 09 17:28:45 crc kubenswrapper[4954]: I1209 17:28:45.896662 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z88n9" Dec 09 17:28:45 crc kubenswrapper[4954]: I1209 17:28:45.942615 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kdt2k" Dec 09 17:28:45 crc kubenswrapper[4954]: I1209 17:28:45.960200 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a4ac-account-create-update-qprph" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.015451 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c84cd" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.018536 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-602a-account-create-update-892sk" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.026900 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.044234 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stwx5\" (UniqueName: \"kubernetes.io/projected/10f8b023-726d-443f-b728-eaa574d45bab-kube-api-access-stwx5\") pod \"10f8b023-726d-443f-b728-eaa574d45bab\" (UID: \"10f8b023-726d-443f-b728-eaa574d45bab\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.044339 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/588cbebb-ba39-4145-9c5e-865f77ee92cb-operator-scripts\") pod \"588cbebb-ba39-4145-9c5e-865f77ee92cb\" (UID: \"588cbebb-ba39-4145-9c5e-865f77ee92cb\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.044382 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10f8b023-726d-443f-b728-eaa574d45bab-operator-scripts\") pod \"10f8b023-726d-443f-b728-eaa574d45bab\" (UID: \"10f8b023-726d-443f-b728-eaa574d45bab\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.044479 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41a8d52b-549b-42d1-91bf-e7e4567affb3-operator-scripts\") pod \"41a8d52b-549b-42d1-91bf-e7e4567affb3\" (UID: \"41a8d52b-549b-42d1-91bf-e7e4567affb3\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.044655 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8gbt\" (UniqueName: \"kubernetes.io/projected/588cbebb-ba39-4145-9c5e-865f77ee92cb-kube-api-access-j8gbt\") pod \"588cbebb-ba39-4145-9c5e-865f77ee92cb\" (UID: \"588cbebb-ba39-4145-9c5e-865f77ee92cb\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.044715 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqqr5\" (UniqueName: \"kubernetes.io/projected/41a8d52b-549b-42d1-91bf-e7e4567affb3-kube-api-access-lqqr5\") pod \"41a8d52b-549b-42d1-91bf-e7e4567affb3\" (UID: \"41a8d52b-549b-42d1-91bf-e7e4567affb3\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.046358 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/588cbebb-ba39-4145-9c5e-865f77ee92cb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "588cbebb-ba39-4145-9c5e-865f77ee92cb" (UID: "588cbebb-ba39-4145-9c5e-865f77ee92cb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.046429 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10f8b023-726d-443f-b728-eaa574d45bab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "10f8b023-726d-443f-b728-eaa574d45bab" (UID: "10f8b023-726d-443f-b728-eaa574d45bab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.046926 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41a8d52b-549b-42d1-91bf-e7e4567affb3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "41a8d52b-549b-42d1-91bf-e7e4567affb3" (UID: "41a8d52b-549b-42d1-91bf-e7e4567affb3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.052985 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41a8d52b-549b-42d1-91bf-e7e4567affb3-kube-api-access-lqqr5" (OuterVolumeSpecName: "kube-api-access-lqqr5") pod "41a8d52b-549b-42d1-91bf-e7e4567affb3" (UID: "41a8d52b-549b-42d1-91bf-e7e4567affb3"). InnerVolumeSpecName "kube-api-access-lqqr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.053968 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/588cbebb-ba39-4145-9c5e-865f77ee92cb-kube-api-access-j8gbt" (OuterVolumeSpecName: "kube-api-access-j8gbt") pod "588cbebb-ba39-4145-9c5e-865f77ee92cb" (UID: "588cbebb-ba39-4145-9c5e-865f77ee92cb"). InnerVolumeSpecName "kube-api-access-j8gbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.054939 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10f8b023-726d-443f-b728-eaa574d45bab-kube-api-access-stwx5" (OuterVolumeSpecName: "kube-api-access-stwx5") pod "10f8b023-726d-443f-b728-eaa574d45bab" (UID: "10f8b023-726d-443f-b728-eaa574d45bab"). InnerVolumeSpecName "kube-api-access-stwx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.147497 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7d295cb-5d01-4014-abbe-efc80d8259f8-operator-scripts\") pod \"d7d295cb-5d01-4014-abbe-efc80d8259f8\" (UID: \"d7d295cb-5d01-4014-abbe-efc80d8259f8\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.147683 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqb4l\" (UniqueName: \"kubernetes.io/projected/1683bbec-ff24-4d14-aed8-cfb49b328355-kube-api-access-vqb4l\") pod \"1683bbec-ff24-4d14-aed8-cfb49b328355\" (UID: \"1683bbec-ff24-4d14-aed8-cfb49b328355\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.147741 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1683bbec-ff24-4d14-aed8-cfb49b328355-operator-scripts\") pod \"1683bbec-ff24-4d14-aed8-cfb49b328355\" (UID: \"1683bbec-ff24-4d14-aed8-cfb49b328355\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.147872 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjgc9\" (UniqueName: \"kubernetes.io/projected/f599c00a-fbf5-46cb-9304-f5554664eb3e-kube-api-access-tjgc9\") pod \"f599c00a-fbf5-46cb-9304-f5554664eb3e\" (UID: \"f599c00a-fbf5-46cb-9304-f5554664eb3e\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.147923 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5vd2\" (UniqueName: \"kubernetes.io/projected/d7d295cb-5d01-4014-abbe-efc80d8259f8-kube-api-access-f5vd2\") pod \"d7d295cb-5d01-4014-abbe-efc80d8259f8\" (UID: \"d7d295cb-5d01-4014-abbe-efc80d8259f8\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.147974 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f599c00a-fbf5-46cb-9304-f5554664eb3e-operator-scripts\") pod \"f599c00a-fbf5-46cb-9304-f5554664eb3e\" (UID: \"f599c00a-fbf5-46cb-9304-f5554664eb3e\") " Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.148394 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d295cb-5d01-4014-abbe-efc80d8259f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7d295cb-5d01-4014-abbe-efc80d8259f8" (UID: "d7d295cb-5d01-4014-abbe-efc80d8259f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.148586 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stwx5\" (UniqueName: \"kubernetes.io/projected/10f8b023-726d-443f-b728-eaa574d45bab-kube-api-access-stwx5\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.148625 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/588cbebb-ba39-4145-9c5e-865f77ee92cb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.148639 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10f8b023-726d-443f-b728-eaa574d45bab-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.148651 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7d295cb-5d01-4014-abbe-efc80d8259f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.148663 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41a8d52b-549b-42d1-91bf-e7e4567affb3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.148677 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8gbt\" (UniqueName: \"kubernetes.io/projected/588cbebb-ba39-4145-9c5e-865f77ee92cb-kube-api-access-j8gbt\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.148689 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqqr5\" (UniqueName: \"kubernetes.io/projected/41a8d52b-549b-42d1-91bf-e7e4567affb3-kube-api-access-lqqr5\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.149055 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f599c00a-fbf5-46cb-9304-f5554664eb3e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f599c00a-fbf5-46cb-9304-f5554664eb3e" (UID: "f599c00a-fbf5-46cb-9304-f5554664eb3e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.149352 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1683bbec-ff24-4d14-aed8-cfb49b328355-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1683bbec-ff24-4d14-aed8-cfb49b328355" (UID: "1683bbec-ff24-4d14-aed8-cfb49b328355"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.154895 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1683bbec-ff24-4d14-aed8-cfb49b328355-kube-api-access-vqb4l" (OuterVolumeSpecName: "kube-api-access-vqb4l") pod "1683bbec-ff24-4d14-aed8-cfb49b328355" (UID: "1683bbec-ff24-4d14-aed8-cfb49b328355"). InnerVolumeSpecName "kube-api-access-vqb4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.154976 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f599c00a-fbf5-46cb-9304-f5554664eb3e-kube-api-access-tjgc9" (OuterVolumeSpecName: "kube-api-access-tjgc9") pod "f599c00a-fbf5-46cb-9304-f5554664eb3e" (UID: "f599c00a-fbf5-46cb-9304-f5554664eb3e"). InnerVolumeSpecName "kube-api-access-tjgc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.155787 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d295cb-5d01-4014-abbe-efc80d8259f8-kube-api-access-f5vd2" (OuterVolumeSpecName: "kube-api-access-f5vd2") pod "d7d295cb-5d01-4014-abbe-efc80d8259f8" (UID: "d7d295cb-5d01-4014-abbe-efc80d8259f8"). InnerVolumeSpecName "kube-api-access-f5vd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.179538 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z88n9" event={"ID":"41a8d52b-549b-42d1-91bf-e7e4567affb3","Type":"ContainerDied","Data":"c7a3ffb103fef31466f1074b75916ad6309e0a728625c5b16259c274971e1f8b"} Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.179966 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7a3ffb103fef31466f1074b75916ad6309e0a728625c5b16259c274971e1f8b" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.179559 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z88n9" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.188378 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" event={"ID":"91fe1cc4-243c-473f-a86a-2a961648027f","Type":"ContainerStarted","Data":"e6ff16410ac7dbd26190e0fdad3be8a81617e6d3683653a874372239b7c33910"} Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.189447 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.197786 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c84cd" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.197820 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-c84cd" event={"ID":"d7d295cb-5d01-4014-abbe-efc80d8259f8","Type":"ContainerDied","Data":"f212cc5e7475c0facaee6067e00af723e6b1e513b72f6564b335bb215823add2"} Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.197874 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f212cc5e7475c0facaee6067e00af723e6b1e513b72f6564b335bb215823add2" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.201876 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-602a-account-create-update-892sk" event={"ID":"1683bbec-ff24-4d14-aed8-cfb49b328355","Type":"ContainerDied","Data":"130e33fa5950599aa47163d6466aed92d8f1e7fbf5371e130c6791d4d2111ca0"} Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.201944 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="130e33fa5950599aa47163d6466aed92d8f1e7fbf5371e130c6791d4d2111ca0" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.202047 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-602a-account-create-update-892sk" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.216833 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kdt2k" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.217192 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kdt2k" event={"ID":"10f8b023-726d-443f-b728-eaa574d45bab","Type":"ContainerDied","Data":"00cb7d8efd62d1f867747365e8abb3fd8e1292d542bf3326e12dbc44c0b37f8f"} Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.217273 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00cb7d8efd62d1f867747365e8abb3fd8e1292d542bf3326e12dbc44c0b37f8f" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.224161 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a4ac-account-create-update-qprph" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.224168 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a4ac-account-create-update-qprph" event={"ID":"588cbebb-ba39-4145-9c5e-865f77ee92cb","Type":"ContainerDied","Data":"79329071ad481ea57a0852de3ac42f2d3cc754b13044c9e48918aed5646b044e"} Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.224219 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79329071ad481ea57a0852de3ac42f2d3cc754b13044c9e48918aed5646b044e" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.224812 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" podStartSLOduration=2.888319808 podStartE2EDuration="6.224784018s" podCreationTimestamp="2025-12-09 17:28:40 +0000 UTC" firstStartedPulling="2025-12-09 17:28:42.139540127 +0000 UTC m=+1918.527713947" lastFinishedPulling="2025-12-09 17:28:45.476004337 +0000 UTC m=+1921.864178157" observedRunningTime="2025-12-09 17:28:46.207186757 +0000 UTC m=+1922.595360587" watchObservedRunningTime="2025-12-09 17:28:46.224784018 +0000 UTC m=+1922.612957838" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.233267 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" event={"ID":"f599c00a-fbf5-46cb-9304-f5554664eb3e","Type":"ContainerDied","Data":"1108407dbd388d6e77dd6c15a950bf1a22e516ef8ac0e5a720539619be9e6243"} Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.233305 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1108407dbd388d6e77dd6c15a950bf1a22e516ef8ac0e5a720539619be9e6243" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.233365 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-bce8-account-create-update-lhxt6" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.240473 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5f58db749d-zq2l7" event={"ID":"8b8a6386-1d71-4193-8d87-303561f4b34a","Type":"ContainerStarted","Data":"dfc03022199afc816925f2558c4b281f17be86e6e5ace7cd4ce0e19c4fb9e67e"} Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.251942 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqb4l\" (UniqueName: \"kubernetes.io/projected/1683bbec-ff24-4d14-aed8-cfb49b328355-kube-api-access-vqb4l\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.251982 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1683bbec-ff24-4d14-aed8-cfb49b328355-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.251995 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjgc9\" (UniqueName: \"kubernetes.io/projected/f599c00a-fbf5-46cb-9304-f5554664eb3e-kube-api-access-tjgc9\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.252007 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5vd2\" (UniqueName: \"kubernetes.io/projected/d7d295cb-5d01-4014-abbe-efc80d8259f8-kube-api-access-f5vd2\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.252021 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f599c00a-fbf5-46cb-9304-f5554664eb3e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:46 crc kubenswrapper[4954]: I1209 17:28:46.276088 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5f58db749d-zq2l7" podStartSLOduration=3.223578403 podStartE2EDuration="6.276057923s" podCreationTimestamp="2025-12-09 17:28:40 +0000 UTC" firstStartedPulling="2025-12-09 17:28:42.397507323 +0000 UTC m=+1918.785681143" lastFinishedPulling="2025-12-09 17:28:45.449986833 +0000 UTC m=+1921.838160663" observedRunningTime="2025-12-09 17:28:46.266986779 +0000 UTC m=+1922.655160619" watchObservedRunningTime="2025-12-09 17:28:46.276057923 +0000 UTC m=+1922.664231743" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.253584 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.925852 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-85d59798d9-p56m8"] Dec 09 17:28:47 crc kubenswrapper[4954]: E1209 17:28:47.926794 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f8b023-726d-443f-b728-eaa574d45bab" containerName="mariadb-database-create" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.926813 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f8b023-726d-443f-b728-eaa574d45bab" containerName="mariadb-database-create" Dec 09 17:28:47 crc kubenswrapper[4954]: E1209 17:28:47.926837 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="588cbebb-ba39-4145-9c5e-865f77ee92cb" containerName="mariadb-account-create-update" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.926845 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="588cbebb-ba39-4145-9c5e-865f77ee92cb" containerName="mariadb-account-create-update" Dec 09 17:28:47 crc kubenswrapper[4954]: E1209 17:28:47.926871 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d295cb-5d01-4014-abbe-efc80d8259f8" containerName="mariadb-database-create" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.926879 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d295cb-5d01-4014-abbe-efc80d8259f8" containerName="mariadb-database-create" Dec 09 17:28:47 crc kubenswrapper[4954]: E1209 17:28:47.926892 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41a8d52b-549b-42d1-91bf-e7e4567affb3" containerName="mariadb-database-create" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.926899 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="41a8d52b-549b-42d1-91bf-e7e4567affb3" containerName="mariadb-database-create" Dec 09 17:28:47 crc kubenswrapper[4954]: E1209 17:28:47.926930 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1683bbec-ff24-4d14-aed8-cfb49b328355" containerName="mariadb-account-create-update" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.926937 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="1683bbec-ff24-4d14-aed8-cfb49b328355" containerName="mariadb-account-create-update" Dec 09 17:28:47 crc kubenswrapper[4954]: E1209 17:28:47.926945 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f599c00a-fbf5-46cb-9304-f5554664eb3e" containerName="mariadb-account-create-update" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.926951 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f599c00a-fbf5-46cb-9304-f5554664eb3e" containerName="mariadb-account-create-update" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.927165 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d295cb-5d01-4014-abbe-efc80d8259f8" containerName="mariadb-database-create" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.927187 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="588cbebb-ba39-4145-9c5e-865f77ee92cb" containerName="mariadb-account-create-update" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.927214 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="41a8d52b-549b-42d1-91bf-e7e4567affb3" containerName="mariadb-database-create" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.927229 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f599c00a-fbf5-46cb-9304-f5554664eb3e" containerName="mariadb-account-create-update" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.927237 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f8b023-726d-443f-b728-eaa574d45bab" containerName="mariadb-database-create" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.927257 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="1683bbec-ff24-4d14-aed8-cfb49b328355" containerName="mariadb-account-create-update" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.928300 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.940368 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-85d59798d9-p56m8"] Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.960367 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6d888884cc-94rdv"] Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.962038 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.992646 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r65r6\" (UniqueName: \"kubernetes.io/projected/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-kube-api-access-r65r6\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.992732 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-combined-ca-bundle\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.992802 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data-custom\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:47 crc kubenswrapper[4954]: I1209 17:28:47.992934 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.005657 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6d888884cc-94rdv"] Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.052332 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-595f8f4f56-t8rrw"] Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.055150 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.095872 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data-custom\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.098872 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.099308 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0a68d8-91d2-4453-8a36-6e4616a88219-config-data\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.099446 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fskk\" (UniqueName: \"kubernetes.io/projected/5d0a68d8-91d2-4453-8a36-6e4616a88219-kube-api-access-9fskk\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.099758 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r65r6\" (UniqueName: \"kubernetes.io/projected/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-kube-api-access-r65r6\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.099861 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d0a68d8-91d2-4453-8a36-6e4616a88219-config-data-custom\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.100021 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-combined-ca-bundle\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.100122 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0a68d8-91d2-4453-8a36-6e4616a88219-combined-ca-bundle\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.101104 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-595f8f4f56-t8rrw"] Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.108510 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-combined-ca-bundle\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.124640 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data-custom\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.124654 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.124685 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r65r6\" (UniqueName: \"kubernetes.io/projected/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-kube-api-access-r65r6\") pod \"heat-cfnapi-85d59798d9-p56m8\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.203039 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgj2b\" (UniqueName: \"kubernetes.io/projected/b5678723-dfca-41a0-ac7b-3a4418060e27-kube-api-access-lgj2b\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.203109 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0a68d8-91d2-4453-8a36-6e4616a88219-config-data\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.203134 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.203153 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fskk\" (UniqueName: \"kubernetes.io/projected/5d0a68d8-91d2-4453-8a36-6e4616a88219-kube-api-access-9fskk\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.203204 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d0a68d8-91d2-4453-8a36-6e4616a88219-config-data-custom\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.203227 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data-custom\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.203275 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0a68d8-91d2-4453-8a36-6e4616a88219-combined-ca-bundle\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.203455 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-combined-ca-bundle\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.209021 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0a68d8-91d2-4453-8a36-6e4616a88219-combined-ca-bundle\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.209157 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0a68d8-91d2-4453-8a36-6e4616a88219-config-data\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.209453 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d0a68d8-91d2-4453-8a36-6e4616a88219-config-data-custom\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.223567 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fskk\" (UniqueName: \"kubernetes.io/projected/5d0a68d8-91d2-4453-8a36-6e4616a88219-kube-api-access-9fskk\") pod \"heat-engine-6d888884cc-94rdv\" (UID: \"5d0a68d8-91d2-4453-8a36-6e4616a88219\") " pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.254379 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.306145 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgj2b\" (UniqueName: \"kubernetes.io/projected/b5678723-dfca-41a0-ac7b-3a4418060e27-kube-api-access-lgj2b\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.306470 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.306666 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data-custom\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.306863 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-combined-ca-bundle\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.311015 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data-custom\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.311486 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-combined-ca-bundle\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.312408 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.317248 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.328252 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgj2b\" (UniqueName: \"kubernetes.io/projected/b5678723-dfca-41a0-ac7b-3a4418060e27-kube-api-access-lgj2b\") pod \"heat-api-595f8f4f56-t8rrw\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.384719 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.781331 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-85d59798d9-p56m8"] Dec 09 17:28:48 crc kubenswrapper[4954]: I1209 17:28:48.919615 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6d888884cc-94rdv"] Dec 09 17:28:49 crc kubenswrapper[4954]: I1209 17:28:49.053670 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-595f8f4f56-t8rrw"] Dec 09 17:28:49 crc kubenswrapper[4954]: I1209 17:28:49.283124 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-595f8f4f56-t8rrw" event={"ID":"b5678723-dfca-41a0-ac7b-3a4418060e27","Type":"ContainerStarted","Data":"f6a6c9841603596d4bb041a8071abec14290eb2298abeaa5a4ff1b9c6ee4ca9e"} Dec 09 17:28:49 crc kubenswrapper[4954]: I1209 17:28:49.290994 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-85d59798d9-p56m8" event={"ID":"ae072c84-04f0-48a3-a9ee-1068a9ae7a34","Type":"ContainerStarted","Data":"5ec98d341e85bc6a4d78945cb62cbed0428b580b0cfb9b64aea6f02b163f8846"} Dec 09 17:28:49 crc kubenswrapper[4954]: I1209 17:28:49.291062 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-85d59798d9-p56m8" event={"ID":"ae072c84-04f0-48a3-a9ee-1068a9ae7a34","Type":"ContainerStarted","Data":"49cc756df164c5840012eba4b031006f4306205a489db72e2bc1dcbdf24ef7df"} Dec 09 17:28:49 crc kubenswrapper[4954]: I1209 17:28:49.291206 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:49 crc kubenswrapper[4954]: I1209 17:28:49.296118 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6d888884cc-94rdv" event={"ID":"5d0a68d8-91d2-4453-8a36-6e4616a88219","Type":"ContainerStarted","Data":"5726179263ececa8c036508c53d15844d7774122765ab5e5cd90df66fcbf2b28"} Dec 09 17:28:49 crc kubenswrapper[4954]: I1209 17:28:49.296213 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6d888884cc-94rdv" event={"ID":"5d0a68d8-91d2-4453-8a36-6e4616a88219","Type":"ContainerStarted","Data":"61ef4278e2482997c44d9c6228748d808d76ce034a95ee64f932f259879cfac8"} Dec 09 17:28:49 crc kubenswrapper[4954]: I1209 17:28:49.296546 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:28:49 crc kubenswrapper[4954]: I1209 17:28:49.324159 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-85d59798d9-p56m8" podStartSLOduration=2.324139515 podStartE2EDuration="2.324139515s" podCreationTimestamp="2025-12-09 17:28:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:49.314717339 +0000 UTC m=+1925.702891159" watchObservedRunningTime="2025-12-09 17:28:49.324139515 +0000 UTC m=+1925.712313335" Dec 09 17:28:49 crc kubenswrapper[4954]: I1209 17:28:49.351658 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6d888884cc-94rdv" podStartSLOduration=2.351630005 podStartE2EDuration="2.351630005s" podCreationTimestamp="2025-12-09 17:28:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:49.33771692 +0000 UTC m=+1925.725890750" watchObservedRunningTime="2025-12-09 17:28:49.351630005 +0000 UTC m=+1925.739803825" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.263276 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5f58db749d-zq2l7"] Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.263698 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-5f58db749d-zq2l7" podUID="8b8a6386-1d71-4193-8d87-303561f4b34a" containerName="heat-api" containerID="cri-o://dfc03022199afc816925f2558c4b281f17be86e6e5ace7cd4ce0e19c4fb9e67e" gracePeriod=60 Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.280136 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7c8bd6d4b7-mg8lb"] Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.280445 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" podUID="91fe1cc4-243c-473f-a86a-2a961648027f" containerName="heat-cfnapi" containerID="cri-o://e6ff16410ac7dbd26190e0fdad3be8a81617e6d3683653a874372239b7c33910" gracePeriod=60 Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.308119 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6744cc5f97-xg7zz"] Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.310667 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.317438 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.317933 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.320395 4954 generic.go:334] "Generic (PLEG): container finished" podID="b5678723-dfca-41a0-ac7b-3a4418060e27" containerID="e05248c7b7f4eea38cb509d3353ebe39ed96f4d0f77d21828377325441431a81" exitCode=1 Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.320960 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-595f8f4f56-t8rrw" event={"ID":"b5678723-dfca-41a0-ac7b-3a4418060e27","Type":"ContainerDied","Data":"e05248c7b7f4eea38cb509d3353ebe39ed96f4d0f77d21828377325441431a81"} Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.324927 4954 scope.go:117] "RemoveContainer" containerID="e05248c7b7f4eea38cb509d3353ebe39ed96f4d0f77d21828377325441431a81" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.333525 4954 generic.go:334] "Generic (PLEG): container finished" podID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" containerID="5ec98d341e85bc6a4d78945cb62cbed0428b580b0cfb9b64aea6f02b163f8846" exitCode=1 Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.333746 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-85d59798d9-p56m8" event={"ID":"ae072c84-04f0-48a3-a9ee-1068a9ae7a34","Type":"ContainerDied","Data":"5ec98d341e85bc6a4d78945cb62cbed0428b580b0cfb9b64aea6f02b163f8846"} Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.334562 4954 scope.go:117] "RemoveContainer" containerID="5ec98d341e85bc6a4d78945cb62cbed0428b580b0cfb9b64aea6f02b163f8846" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.348670 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7776d8d5fb-j42q6"] Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.351720 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.371523 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.371892 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.378709 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6744cc5f97-xg7zz"] Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.404371 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7776d8d5fb-j42q6"] Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.415027 4954 generic.go:334] "Generic (PLEG): container finished" podID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerID="da7e20528413ebb59131126e1769f5a793003b872b669e19c22f86bc40727556" exitCode=0 Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.415628 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40981fe-a91e-45ef-a39c-ef449c19c751","Type":"ContainerDied","Data":"da7e20528413ebb59131126e1769f5a793003b872b669e19c22f86bc40727556"} Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477108 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-combined-ca-bundle\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477168 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-config-data-custom\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477212 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq9qn\" (UniqueName: \"kubernetes.io/projected/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-kube-api-access-rq9qn\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477285 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-internal-tls-certs\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477342 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-config-data\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477386 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgd5w\" (UniqueName: \"kubernetes.io/projected/1241e34b-a815-4395-8c78-16b79f6e8ecb-kube-api-access-zgd5w\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477418 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-config-data\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477448 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-internal-tls-certs\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477561 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-public-tls-certs\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477588 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-public-tls-certs\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477640 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-config-data-custom\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.477680 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-combined-ca-bundle\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.582510 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-config-data-custom\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.582612 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq9qn\" (UniqueName: \"kubernetes.io/projected/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-kube-api-access-rq9qn\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.582658 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-internal-tls-certs\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.582728 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-config-data\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.582775 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgd5w\" (UniqueName: \"kubernetes.io/projected/1241e34b-a815-4395-8c78-16b79f6e8ecb-kube-api-access-zgd5w\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.582800 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-config-data\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.582845 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-internal-tls-certs\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.582945 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-public-tls-certs\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.582966 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-public-tls-certs\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.583003 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-config-data-custom\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.583032 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-combined-ca-bundle\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.583167 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-combined-ca-bundle\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.597697 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-combined-ca-bundle\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.599962 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-config-data-custom\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.603703 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-config-data\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.605207 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-config-data\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.605858 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-internal-tls-certs\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.612467 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-config-data-custom\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.617544 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-public-tls-certs\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.621306 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-public-tls-certs\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.622343 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-internal-tls-certs\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.626508 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1241e34b-a815-4395-8c78-16b79f6e8ecb-combined-ca-bundle\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.643493 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq9qn\" (UniqueName: \"kubernetes.io/projected/03ea0f62-7526-4cc9-905a-cfd0d5f791e0-kube-api-access-rq9qn\") pod \"heat-api-6744cc5f97-xg7zz\" (UID: \"03ea0f62-7526-4cc9-905a-cfd0d5f791e0\") " pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.647011 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgd5w\" (UniqueName: \"kubernetes.io/projected/1241e34b-a815-4395-8c78-16b79f6e8ecb-kube-api-access-zgd5w\") pod \"heat-cfnapi-7776d8d5fb-j42q6\" (UID: \"1241e34b-a815-4395-8c78-16b79f6e8ecb\") " pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.777706 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.805756 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.963191 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.984013 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lffjc"] Dec 09 17:28:50 crc kubenswrapper[4954]: I1209 17:28:50.984291 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" podUID="d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" containerName="dnsmasq-dns" containerID="cri-o://f14e6b3ced7be04a32d22b82a3f3acd5a88b7b5a16f4970ad96ca1e6bb5ca326" gracePeriod=10 Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.038704 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ln5zl"] Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.040573 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.044040 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-zj68k" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.045360 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.045500 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.086677 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ln5zl"] Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.116112 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-config-data\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.116641 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.116692 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-scripts\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.116908 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78fx2\" (UniqueName: \"kubernetes.io/projected/24d42cd5-dc37-4675-87b9-91dcb7391eed-kube-api-access-78fx2\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.219278 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-config-data\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.219355 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.219388 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-scripts\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.219531 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78fx2\" (UniqueName: \"kubernetes.io/projected/24d42cd5-dc37-4675-87b9-91dcb7391eed-kube-api-access-78fx2\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.229542 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-config-data\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.236759 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-scripts\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.238622 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.253732 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" podUID="91fe1cc4-243c-473f-a86a-2a961648027f" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.207:8000/healthcheck\": read tcp 10.217.0.2:40712->10.217.0.207:8000: read: connection reset by peer" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.254920 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78fx2\" (UniqueName: \"kubernetes.io/projected/24d42cd5-dc37-4675-87b9-91dcb7391eed-kube-api-access-78fx2\") pod \"nova-cell0-conductor-db-sync-ln5zl\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.278902 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5f58db749d-zq2l7" podUID="8b8a6386-1d71-4193-8d87-303561f4b34a" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.209:8004/healthcheck\": read tcp 10.217.0.2:52896->10.217.0.209:8004: read: connection reset by peer" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.434889 4954 generic.go:334] "Generic (PLEG): container finished" podID="91fe1cc4-243c-473f-a86a-2a961648027f" containerID="e6ff16410ac7dbd26190e0fdad3be8a81617e6d3683653a874372239b7c33910" exitCode=0 Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.434950 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" event={"ID":"91fe1cc4-243c-473f-a86a-2a961648027f","Type":"ContainerDied","Data":"e6ff16410ac7dbd26190e0fdad3be8a81617e6d3683653a874372239b7c33910"} Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.438675 4954 generic.go:334] "Generic (PLEG): container finished" podID="8b8a6386-1d71-4193-8d87-303561f4b34a" containerID="dfc03022199afc816925f2558c4b281f17be86e6e5ace7cd4ce0e19c4fb9e67e" exitCode=0 Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.438752 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5f58db749d-zq2l7" event={"ID":"8b8a6386-1d71-4193-8d87-303561f4b34a","Type":"ContainerDied","Data":"dfc03022199afc816925f2558c4b281f17be86e6e5ace7cd4ce0e19c4fb9e67e"} Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.453734 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-595f8f4f56-t8rrw" event={"ID":"b5678723-dfca-41a0-ac7b-3a4418060e27","Type":"ContainerStarted","Data":"96d102e3d8339929771b718fa21c22c75ce3db838a66e2741856528d82cfd0b7"} Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.454247 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.462225 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.464730 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-85d59798d9-p56m8" event={"ID":"ae072c84-04f0-48a3-a9ee-1068a9ae7a34","Type":"ContainerStarted","Data":"9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8"} Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.465780 4954 scope.go:117] "RemoveContainer" containerID="9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8" Dec 09 17:28:51 crc kubenswrapper[4954]: E1209 17:28:51.466037 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-85d59798d9-p56m8_openstack(ae072c84-04f0-48a3-a9ee-1068a9ae7a34)\"" pod="openstack/heat-cfnapi-85d59798d9-p56m8" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.471313 4954 generic.go:334] "Generic (PLEG): container finished" podID="d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" containerID="f14e6b3ced7be04a32d22b82a3f3acd5a88b7b5a16f4970ad96ca1e6bb5ca326" exitCode=0 Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.471357 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" event={"ID":"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be","Type":"ContainerDied","Data":"f14e6b3ced7be04a32d22b82a3f3acd5a88b7b5a16f4970ad96ca1e6bb5ca326"} Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.535633 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-595f8f4f56-t8rrw" podStartSLOduration=4.535606636 podStartE2EDuration="4.535606636s" podCreationTimestamp="2025-12-09 17:28:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:51.48304065 +0000 UTC m=+1927.871214470" watchObservedRunningTime="2025-12-09 17:28:51.535606636 +0000 UTC m=+1927.923780456" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.538395 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7776d8d5fb-j42q6"] Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.851466 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.964986 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwdhw\" (UniqueName: \"kubernetes.io/projected/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-kube-api-access-dwdhw\") pod \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.965081 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-config\") pod \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.967469 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-svc\") pod \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.968123 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-sb\") pod \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.968156 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-swift-storage-0\") pod \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.968187 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-nb\") pod \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\" (UID: \"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be\") " Dec 09 17:28:51 crc kubenswrapper[4954]: I1209 17:28:51.985855 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-kube-api-access-dwdhw" (OuterVolumeSpecName: "kube-api-access-dwdhw") pod "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" (UID: "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be"). InnerVolumeSpecName "kube-api-access-dwdhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.053696 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6744cc5f97-xg7zz"] Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.074911 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwdhw\" (UniqueName: \"kubernetes.io/projected/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-kube-api-access-dwdhw\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.080542 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" (UID: "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.121371 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:28:52 crc kubenswrapper[4954]: E1209 17:28:52.121761 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.145415 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" (UID: "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.157235 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-config" (OuterVolumeSpecName: "config") pod "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" (UID: "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.192372 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.192409 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.194950 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.251254 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" (UID: "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.258309 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" (UID: "d002f0cb-84e5-4bd3-a5f9-4ac71789e0be"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.297882 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.297924 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.303978 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.306634 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ln5zl"] Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.307927 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.400050 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data-custom\") pod \"8b8a6386-1d71-4193-8d87-303561f4b34a\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.400132 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data-custom\") pod \"91fe1cc4-243c-473f-a86a-2a961648027f\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.400339 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9whpx\" (UniqueName: \"kubernetes.io/projected/91fe1cc4-243c-473f-a86a-2a961648027f-kube-api-access-9whpx\") pod \"91fe1cc4-243c-473f-a86a-2a961648027f\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.400372 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-combined-ca-bundle\") pod \"8b8a6386-1d71-4193-8d87-303561f4b34a\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.400394 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data\") pod \"91fe1cc4-243c-473f-a86a-2a961648027f\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.400414 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data\") pod \"8b8a6386-1d71-4193-8d87-303561f4b34a\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.400457 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-combined-ca-bundle\") pod \"91fe1cc4-243c-473f-a86a-2a961648027f\" (UID: \"91fe1cc4-243c-473f-a86a-2a961648027f\") " Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.400505 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvz94\" (UniqueName: \"kubernetes.io/projected/8b8a6386-1d71-4193-8d87-303561f4b34a-kube-api-access-lvz94\") pod \"8b8a6386-1d71-4193-8d87-303561f4b34a\" (UID: \"8b8a6386-1d71-4193-8d87-303561f4b34a\") " Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.408202 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b8a6386-1d71-4193-8d87-303561f4b34a-kube-api-access-lvz94" (OuterVolumeSpecName: "kube-api-access-lvz94") pod "8b8a6386-1d71-4193-8d87-303561f4b34a" (UID: "8b8a6386-1d71-4193-8d87-303561f4b34a"). InnerVolumeSpecName "kube-api-access-lvz94". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.414497 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "91fe1cc4-243c-473f-a86a-2a961648027f" (UID: "91fe1cc4-243c-473f-a86a-2a961648027f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.416287 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91fe1cc4-243c-473f-a86a-2a961648027f-kube-api-access-9whpx" (OuterVolumeSpecName: "kube-api-access-9whpx") pod "91fe1cc4-243c-473f-a86a-2a961648027f" (UID: "91fe1cc4-243c-473f-a86a-2a961648027f"). InnerVolumeSpecName "kube-api-access-9whpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.447085 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8b8a6386-1d71-4193-8d87-303561f4b34a" (UID: "8b8a6386-1d71-4193-8d87-303561f4b34a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.465355 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91fe1cc4-243c-473f-a86a-2a961648027f" (UID: "91fe1cc4-243c-473f-a86a-2a961648027f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.517981 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.518013 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvz94\" (UniqueName: \"kubernetes.io/projected/8b8a6386-1d71-4193-8d87-303561f4b34a-kube-api-access-lvz94\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.518025 4954 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.518034 4954 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.518043 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9whpx\" (UniqueName: \"kubernetes.io/projected/91fe1cc4-243c-473f-a86a-2a961648027f-kube-api-access-9whpx\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.522088 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" event={"ID":"d002f0cb-84e5-4bd3-a5f9-4ac71789e0be","Type":"ContainerDied","Data":"1c81aad3430743b8ebe79fcc0b892cf6f12b0845bc9d379cabbf5378db4647b7"} Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.522177 4954 scope.go:117] "RemoveContainer" containerID="f14e6b3ced7be04a32d22b82a3f3acd5a88b7b5a16f4970ad96ca1e6bb5ca326" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.522717 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-lffjc" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.537067 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" event={"ID":"91fe1cc4-243c-473f-a86a-2a961648027f","Type":"ContainerDied","Data":"6ddfd65caf3219c2dfb444d2eb22bc682fc0c6b9ce693d51da113bdebd3768f4"} Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.537193 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7c8bd6d4b7-mg8lb" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.550680 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b8a6386-1d71-4193-8d87-303561f4b34a" (UID: "8b8a6386-1d71-4193-8d87-303561f4b34a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.559192 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5f58db749d-zq2l7" event={"ID":"8b8a6386-1d71-4193-8d87-303561f4b34a","Type":"ContainerDied","Data":"4311ac47319841ee55591d6a90ab452ac77a005742c5819b790ceb70942d054a"} Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.559311 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5f58db749d-zq2l7" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.575184 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6744cc5f97-xg7zz" event={"ID":"03ea0f62-7526-4cc9-905a-cfd0d5f791e0","Type":"ContainerStarted","Data":"019ea90b818f3660be73dea566dabf5124bf1cf408532fc467bcf9ab8b69ab2a"} Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.601290 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data" (OuterVolumeSpecName: "config-data") pod "91fe1cc4-243c-473f-a86a-2a961648027f" (UID: "91fe1cc4-243c-473f-a86a-2a961648027f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.606748 4954 generic.go:334] "Generic (PLEG): container finished" podID="b5678723-dfca-41a0-ac7b-3a4418060e27" containerID="96d102e3d8339929771b718fa21c22c75ce3db838a66e2741856528d82cfd0b7" exitCode=1 Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.606778 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-595f8f4f56-t8rrw" event={"ID":"b5678723-dfca-41a0-ac7b-3a4418060e27","Type":"ContainerDied","Data":"96d102e3d8339929771b718fa21c22c75ce3db838a66e2741856528d82cfd0b7"} Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.607635 4954 scope.go:117] "RemoveContainer" containerID="96d102e3d8339929771b718fa21c22c75ce3db838a66e2741856528d82cfd0b7" Dec 09 17:28:52 crc kubenswrapper[4954]: E1209 17:28:52.608062 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-595f8f4f56-t8rrw_openstack(b5678723-dfca-41a0-ac7b-3a4418060e27)\"" pod="openstack/heat-api-595f8f4f56-t8rrw" podUID="b5678723-dfca-41a0-ac7b-3a4418060e27" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.614219 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data" (OuterVolumeSpecName: "config-data") pod "8b8a6386-1d71-4193-8d87-303561f4b34a" (UID: "8b8a6386-1d71-4193-8d87-303561f4b34a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.620249 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.620286 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91fe1cc4-243c-473f-a86a-2a961648027f-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.620298 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b8a6386-1d71-4193-8d87-303561f4b34a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.623995 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" event={"ID":"1241e34b-a815-4395-8c78-16b79f6e8ecb","Type":"ContainerStarted","Data":"cea4eb42ad3a825c96c3ee7466eee663997216df9f3f19918d5a91245dd826f9"} Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.624062 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" event={"ID":"1241e34b-a815-4395-8c78-16b79f6e8ecb","Type":"ContainerStarted","Data":"8c585b1f59f837fae6b9004cbf7153adb1d8ac59437ab54dffc6d2bf8a0457a8"} Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.624380 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.642042 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ln5zl" event={"ID":"24d42cd5-dc37-4675-87b9-91dcb7391eed","Type":"ContainerStarted","Data":"3bb6185f96517e853a1f4c4b427d57d83335534a30dde20f28a396fe8ddbe363"} Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.654520 4954 generic.go:334] "Generic (PLEG): container finished" podID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" containerID="9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8" exitCode=1 Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.654586 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-85d59798d9-p56m8" event={"ID":"ae072c84-04f0-48a3-a9ee-1068a9ae7a34","Type":"ContainerDied","Data":"9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8"} Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.655490 4954 scope.go:117] "RemoveContainer" containerID="9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8" Dec 09 17:28:52 crc kubenswrapper[4954]: E1209 17:28:52.655775 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-85d59798d9-p56m8_openstack(ae072c84-04f0-48a3-a9ee-1068a9ae7a34)\"" pod="openstack/heat-cfnapi-85d59798d9-p56m8" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.699046 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" podStartSLOduration=2.699016296 podStartE2EDuration="2.699016296s" podCreationTimestamp="2025-12-09 17:28:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:52.65408293 +0000 UTC m=+1929.042256750" watchObservedRunningTime="2025-12-09 17:28:52.699016296 +0000 UTC m=+1929.087190116" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.796811 4954 scope.go:117] "RemoveContainer" containerID="e3a76c6136ded675b2709fb350f57424cb4b0ab3524a287e16f7ffe4c028bf9a" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.815686 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lffjc"] Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.818322 4954 scope.go:117] "RemoveContainer" containerID="e6ff16410ac7dbd26190e0fdad3be8a81617e6d3683653a874372239b7c33910" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.827682 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-lffjc"] Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.879760 4954 scope.go:117] "RemoveContainer" containerID="dfc03022199afc816925f2558c4b281f17be86e6e5ace7cd4ce0e19c4fb9e67e" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.893461 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7c8bd6d4b7-mg8lb"] Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.903010 4954 scope.go:117] "RemoveContainer" containerID="e05248c7b7f4eea38cb509d3353ebe39ed96f4d0f77d21828377325441431a81" Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.922245 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7c8bd6d4b7-mg8lb"] Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.953011 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5f58db749d-zq2l7"] Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.978510 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5f58db749d-zq2l7"] Dec 09 17:28:52 crc kubenswrapper[4954]: I1209 17:28:52.985556 4954 scope.go:117] "RemoveContainer" containerID="5ec98d341e85bc6a4d78945cb62cbed0428b580b0cfb9b64aea6f02b163f8846" Dec 09 17:28:53 crc kubenswrapper[4954]: I1209 17:28:53.259528 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:53 crc kubenswrapper[4954]: I1209 17:28:53.259616 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:28:53 crc kubenswrapper[4954]: I1209 17:28:53.385910 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:28:53 crc kubenswrapper[4954]: I1209 17:28:53.670372 4954 scope.go:117] "RemoveContainer" containerID="9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8" Dec 09 17:28:53 crc kubenswrapper[4954]: E1209 17:28:53.671081 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-85d59798d9-p56m8_openstack(ae072c84-04f0-48a3-a9ee-1068a9ae7a34)\"" pod="openstack/heat-cfnapi-85d59798d9-p56m8" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" Dec 09 17:28:53 crc kubenswrapper[4954]: I1209 17:28:53.680719 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6744cc5f97-xg7zz" event={"ID":"03ea0f62-7526-4cc9-905a-cfd0d5f791e0","Type":"ContainerStarted","Data":"6a7d4130025fc733c0f727531b6a45d4874bb385356833744e79965eb4f0f132"} Dec 09 17:28:53 crc kubenswrapper[4954]: I1209 17:28:53.680859 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:28:53 crc kubenswrapper[4954]: I1209 17:28:53.695006 4954 scope.go:117] "RemoveContainer" containerID="96d102e3d8339929771b718fa21c22c75ce3db838a66e2741856528d82cfd0b7" Dec 09 17:28:53 crc kubenswrapper[4954]: E1209 17:28:53.695302 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-595f8f4f56-t8rrw_openstack(b5678723-dfca-41a0-ac7b-3a4418060e27)\"" pod="openstack/heat-api-595f8f4f56-t8rrw" podUID="b5678723-dfca-41a0-ac7b-3a4418060e27" Dec 09 17:28:53 crc kubenswrapper[4954]: I1209 17:28:53.707804 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6744cc5f97-xg7zz" podStartSLOduration=3.707766606 podStartE2EDuration="3.707766606s" podCreationTimestamp="2025-12-09 17:28:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:28:53.701880712 +0000 UTC m=+1930.090054552" watchObservedRunningTime="2025-12-09 17:28:53.707766606 +0000 UTC m=+1930.095940426" Dec 09 17:28:54 crc kubenswrapper[4954]: I1209 17:28:54.136825 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b8a6386-1d71-4193-8d87-303561f4b34a" path="/var/lib/kubelet/pods/8b8a6386-1d71-4193-8d87-303561f4b34a/volumes" Dec 09 17:28:54 crc kubenswrapper[4954]: I1209 17:28:54.137469 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91fe1cc4-243c-473f-a86a-2a961648027f" path="/var/lib/kubelet/pods/91fe1cc4-243c-473f-a86a-2a961648027f/volumes" Dec 09 17:28:54 crc kubenswrapper[4954]: I1209 17:28:54.138010 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" path="/var/lib/kubelet/pods/d002f0cb-84e5-4bd3-a5f9-4ac71789e0be/volumes" Dec 09 17:28:54 crc kubenswrapper[4954]: I1209 17:28:54.706673 4954 scope.go:117] "RemoveContainer" containerID="96d102e3d8339929771b718fa21c22c75ce3db838a66e2741856528d82cfd0b7" Dec 09 17:28:54 crc kubenswrapper[4954]: E1209 17:28:54.706928 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-595f8f4f56-t8rrw_openstack(b5678723-dfca-41a0-ac7b-3a4418060e27)\"" pod="openstack/heat-api-595f8f4f56-t8rrw" podUID="b5678723-dfca-41a0-ac7b-3a4418060e27" Dec 09 17:28:54 crc kubenswrapper[4954]: I1209 17:28:54.707088 4954 scope.go:117] "RemoveContainer" containerID="9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8" Dec 09 17:28:54 crc kubenswrapper[4954]: E1209 17:28:54.707388 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-85d59798d9-p56m8_openstack(ae072c84-04f0-48a3-a9ee-1068a9ae7a34)\"" pod="openstack/heat-cfnapi-85d59798d9-p56m8" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" Dec 09 17:29:00 crc kubenswrapper[4954]: E1209 17:29:00.359176 4954 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/edf8d09de9de6666eaeb50f64e3d93bbc9063675a4f25a00ace427cdd21a4db9/diff" to get inode usage: stat /var/lib/containers/storage/overlay/edf8d09de9de6666eaeb50f64e3d93bbc9063675a4f25a00ace427cdd21a4db9/diff: no such file or directory, extraDiskErr: Dec 09 17:29:00 crc kubenswrapper[4954]: I1209 17:29:00.639729 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:29:01 crc kubenswrapper[4954]: E1209 17:29:01.709401 4954 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/1522541dd996a71035b410c98080d2e2df2351f0cea0dc8e43090d05ca82a78a/diff" to get inode usage: stat /var/lib/containers/storage/overlay/1522541dd996a71035b410c98080d2e2df2351f0cea0dc8e43090d05ca82a78a/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-5c9776ccc5-lffjc_d002f0cb-84e5-4bd3-a5f9-4ac71789e0be/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-5c9776ccc5-lffjc_d002f0cb-84e5-4bd3-a5f9-4ac71789e0be/dnsmasq-dns/0.log: no such file or directory Dec 09 17:29:01 crc kubenswrapper[4954]: I1209 17:29:01.813554 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ln5zl" event={"ID":"24d42cd5-dc37-4675-87b9-91dcb7391eed","Type":"ContainerStarted","Data":"289fc5c5a373dc0fbb22475c3925fb11dc75b13b2f17fb7b28cdf313f453841a"} Dec 09 17:29:01 crc kubenswrapper[4954]: I1209 17:29:01.845029 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-ln5zl" podStartSLOduration=3.637777706 podStartE2EDuration="11.844997177s" podCreationTimestamp="2025-12-09 17:28:50 +0000 UTC" firstStartedPulling="2025-12-09 17:28:52.351503318 +0000 UTC m=+1928.739677138" lastFinishedPulling="2025-12-09 17:29:00.558722789 +0000 UTC m=+1936.946896609" observedRunningTime="2025-12-09 17:29:01.832844566 +0000 UTC m=+1938.221018406" watchObservedRunningTime="2025-12-09 17:29:01.844997177 +0000 UTC m=+1938.233170997" Dec 09 17:29:02 crc kubenswrapper[4954]: I1209 17:29:02.684044 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6744cc5f97-xg7zz" Dec 09 17:29:02 crc kubenswrapper[4954]: I1209 17:29:02.762203 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-595f8f4f56-t8rrw"] Dec 09 17:29:02 crc kubenswrapper[4954]: I1209 17:29:02.923336 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7776d8d5fb-j42q6" Dec 09 17:29:02 crc kubenswrapper[4954]: I1209 17:29:02.998320 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-85d59798d9-p56m8"] Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.259206 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.394457 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data\") pod \"b5678723-dfca-41a0-ac7b-3a4418060e27\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.394626 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgj2b\" (UniqueName: \"kubernetes.io/projected/b5678723-dfca-41a0-ac7b-3a4418060e27-kube-api-access-lgj2b\") pod \"b5678723-dfca-41a0-ac7b-3a4418060e27\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.394654 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data-custom\") pod \"b5678723-dfca-41a0-ac7b-3a4418060e27\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.394723 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-combined-ca-bundle\") pod \"b5678723-dfca-41a0-ac7b-3a4418060e27\" (UID: \"b5678723-dfca-41a0-ac7b-3a4418060e27\") " Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.419969 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b5678723-dfca-41a0-ac7b-3a4418060e27" (UID: "b5678723-dfca-41a0-ac7b-3a4418060e27"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.426799 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5678723-dfca-41a0-ac7b-3a4418060e27-kube-api-access-lgj2b" (OuterVolumeSpecName: "kube-api-access-lgj2b") pod "b5678723-dfca-41a0-ac7b-3a4418060e27" (UID: "b5678723-dfca-41a0-ac7b-3a4418060e27"). InnerVolumeSpecName "kube-api-access-lgj2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.472868 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data" (OuterVolumeSpecName: "config-data") pod "b5678723-dfca-41a0-ac7b-3a4418060e27" (UID: "b5678723-dfca-41a0-ac7b-3a4418060e27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.505328 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.505557 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgj2b\" (UniqueName: \"kubernetes.io/projected/b5678723-dfca-41a0-ac7b-3a4418060e27-kube-api-access-lgj2b\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.505702 4954 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.506661 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5678723-dfca-41a0-ac7b-3a4418060e27" (UID: "b5678723-dfca-41a0-ac7b-3a4418060e27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.607766 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5678723-dfca-41a0-ac7b-3a4418060e27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.610528 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.709567 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data\") pod \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.709816 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r65r6\" (UniqueName: \"kubernetes.io/projected/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-kube-api-access-r65r6\") pod \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.709898 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-combined-ca-bundle\") pod \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.710035 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data-custom\") pod \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\" (UID: \"ae072c84-04f0-48a3-a9ee-1068a9ae7a34\") " Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.716072 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-kube-api-access-r65r6" (OuterVolumeSpecName: "kube-api-access-r65r6") pod "ae072c84-04f0-48a3-a9ee-1068a9ae7a34" (UID: "ae072c84-04f0-48a3-a9ee-1068a9ae7a34"). InnerVolumeSpecName "kube-api-access-r65r6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.716284 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ae072c84-04f0-48a3-a9ee-1068a9ae7a34" (UID: "ae072c84-04f0-48a3-a9ee-1068a9ae7a34"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.744611 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae072c84-04f0-48a3-a9ee-1068a9ae7a34" (UID: "ae072c84-04f0-48a3-a9ee-1068a9ae7a34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.775687 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data" (OuterVolumeSpecName: "config-data") pod "ae072c84-04f0-48a3-a9ee-1068a9ae7a34" (UID: "ae072c84-04f0-48a3-a9ee-1068a9ae7a34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.813131 4954 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.813177 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.813191 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r65r6\" (UniqueName: \"kubernetes.io/projected/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-kube-api-access-r65r6\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.813206 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae072c84-04f0-48a3-a9ee-1068a9ae7a34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.849449 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-595f8f4f56-t8rrw" event={"ID":"b5678723-dfca-41a0-ac7b-3a4418060e27","Type":"ContainerDied","Data":"f6a6c9841603596d4bb041a8071abec14290eb2298abeaa5a4ff1b9c6ee4ca9e"} Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.849540 4954 scope.go:117] "RemoveContainer" containerID="96d102e3d8339929771b718fa21c22c75ce3db838a66e2741856528d82cfd0b7" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.849538 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-595f8f4f56-t8rrw" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.853121 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-85d59798d9-p56m8" event={"ID":"ae072c84-04f0-48a3-a9ee-1068a9ae7a34","Type":"ContainerDied","Data":"49cc756df164c5840012eba4b031006f4306205a489db72e2bc1dcbdf24ef7df"} Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.853224 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-85d59798d9-p56m8" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.889956 4954 scope.go:117] "RemoveContainer" containerID="9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8" Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.908037 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-85d59798d9-p56m8"] Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.939061 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-85d59798d9-p56m8"] Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.960631 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-595f8f4f56-t8rrw"] Dec 09 17:29:03 crc kubenswrapper[4954]: I1209 17:29:03.982457 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-595f8f4f56-t8rrw"] Dec 09 17:29:04 crc kubenswrapper[4954]: I1209 17:29:04.134832 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" path="/var/lib/kubelet/pods/ae072c84-04f0-48a3-a9ee-1068a9ae7a34/volumes" Dec 09 17:29:04 crc kubenswrapper[4954]: I1209 17:29:04.135419 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5678723-dfca-41a0-ac7b-3a4418060e27" path="/var/lib/kubelet/pods/b5678723-dfca-41a0-ac7b-3a4418060e27/volumes" Dec 09 17:29:04 crc kubenswrapper[4954]: I1209 17:29:04.932024 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 17:29:05 crc kubenswrapper[4954]: I1209 17:29:05.120344 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:29:05 crc kubenswrapper[4954]: E1209 17:29:05.120837 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:29:08 crc kubenswrapper[4954]: I1209 17:29:08.356277 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6d888884cc-94rdv" Dec 09 17:29:08 crc kubenswrapper[4954]: I1209 17:29:08.433353 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7b79757c4c-zrc6v"] Dec 09 17:29:08 crc kubenswrapper[4954]: I1209 17:29:08.434200 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-7b79757c4c-zrc6v" podUID="cfe2f3db-ef77-459d-8ee8-e0df346b5beb" containerName="heat-engine" containerID="cri-o://63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e" gracePeriod=60 Dec 09 17:29:10 crc kubenswrapper[4954]: E1209 17:29:10.604107 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 17:29:10 crc kubenswrapper[4954]: E1209 17:29:10.606884 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 17:29:10 crc kubenswrapper[4954]: E1209 17:29:10.608360 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 09 17:29:10 crc kubenswrapper[4954]: E1209 17:29:10.608435 4954 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7b79757c4c-zrc6v" podUID="cfe2f3db-ef77-459d-8ee8-e0df346b5beb" containerName="heat-engine" Dec 09 17:29:10 crc kubenswrapper[4954]: W1209 17:29:10.832102 4954 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5678723_dfca_41a0_ac7b_3a4418060e27.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5678723_dfca_41a0_ac7b_3a4418060e27.slice: no such file or directory Dec 09 17:29:10 crc kubenswrapper[4954]: W1209 17:29:10.832175 4954 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice/crio-49cc756df164c5840012eba4b031006f4306205a489db72e2bc1dcbdf24ef7df": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice/crio-49cc756df164c5840012eba4b031006f4306205a489db72e2bc1dcbdf24ef7df: no such file or directory Dec 09 17:29:10 crc kubenswrapper[4954]: W1209 17:29:10.832189 4954 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice/crio-conmon-5ec98d341e85bc6a4d78945cb62cbed0428b580b0cfb9b64aea6f02b163f8846.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice/crio-conmon-5ec98d341e85bc6a4d78945cb62cbed0428b580b0cfb9b64aea6f02b163f8846.scope: no such file or directory Dec 09 17:29:10 crc kubenswrapper[4954]: W1209 17:29:10.832203 4954 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice/crio-5ec98d341e85bc6a4d78945cb62cbed0428b580b0cfb9b64aea6f02b163f8846.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice/crio-5ec98d341e85bc6a4d78945cb62cbed0428b580b0cfb9b64aea6f02b163f8846.scope: no such file or directory Dec 09 17:29:10 crc kubenswrapper[4954]: W1209 17:29:10.837847 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91fe1cc4_243c_473f_a86a_2a961648027f.slice/crio-e6ff16410ac7dbd26190e0fdad3be8a81617e6d3683653a874372239b7c33910.scope WatchSource:0}: Error finding container e6ff16410ac7dbd26190e0fdad3be8a81617e6d3683653a874372239b7c33910: Status 404 returned error can't find the container with id e6ff16410ac7dbd26190e0fdad3be8a81617e6d3683653a874372239b7c33910 Dec 09 17:29:10 crc kubenswrapper[4954]: W1209 17:29:10.838438 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b8a6386_1d71_4193_8d87_303561f4b34a.slice/crio-dfc03022199afc816925f2558c4b281f17be86e6e5ace7cd4ce0e19c4fb9e67e.scope WatchSource:0}: Error finding container dfc03022199afc816925f2558c4b281f17be86e6e5ace7cd4ce0e19c4fb9e67e: Status 404 returned error can't find the container with id dfc03022199afc816925f2558c4b281f17be86e6e5ace7cd4ce0e19c4fb9e67e Dec 09 17:29:10 crc kubenswrapper[4954]: W1209 17:29:10.841878 4954 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice/crio-conmon-9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice/crio-conmon-9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8.scope: no such file or directory Dec 09 17:29:10 crc kubenswrapper[4954]: W1209 17:29:10.841943 4954 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice/crio-9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice/crio-9adaffae23a06c0132643b730e13f4ae156ba5a24e1fb0d2b21eda2f49db06b8.scope: no such file or directory Dec 09 17:29:10 crc kubenswrapper[4954]: I1209 17:29:10.950145 4954 generic.go:334] "Generic (PLEG): container finished" podID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerID="109b477bde6f46241718782f183745f21cb35376ea49165b7bd443649b0f5566" exitCode=137 Dec 09 17:29:10 crc kubenswrapper[4954]: I1209 17:29:10.950199 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40981fe-a91e-45ef-a39c-ef449c19c751","Type":"ContainerDied","Data":"109b477bde6f46241718782f183745f21cb35376ea49165b7bd443649b0f5566"} Dec 09 17:29:11 crc kubenswrapper[4954]: E1209 17:29:11.249047 4954 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91fe1cc4_243c_473f_a86a_2a961648027f.slice/crio-6ddfd65caf3219c2dfb444d2eb22bc682fc0c6b9ce693d51da113bdebd3768f4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd002f0cb_84e5_4bd3_a5f9_4ac71789e0be.slice/crio-f14e6b3ced7be04a32d22b82a3f3acd5a88b7b5a16f4970ad96ca1e6bb5ca326.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd002f0cb_84e5_4bd3_a5f9_4ac71789e0be.slice/crio-1c81aad3430743b8ebe79fcc0b892cf6f12b0845bc9d379cabbf5378db4647b7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae072c84_04f0_48a3_a9ee_1068a9ae7a34.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b8a6386_1d71_4193_8d87_303561f4b34a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91fe1cc4_243c_473f_a86a_2a961648027f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf40981fe_a91e_45ef_a39c_ef449c19c751.slice/crio-conmon-109b477bde6f46241718782f183745f21cb35376ea49165b7bd443649b0f5566.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd002f0cb_84e5_4bd3_a5f9_4ac71789e0be.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b8a6386_1d71_4193_8d87_303561f4b34a.slice/crio-conmon-dfc03022199afc816925f2558c4b281f17be86e6e5ace7cd4ce0e19c4fb9e67e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b8a6386_1d71_4193_8d87_303561f4b34a.slice/crio-4311ac47319841ee55591d6a90ab452ac77a005742c5819b790ceb70942d054a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91fe1cc4_243c_473f_a86a_2a961648027f.slice/crio-conmon-e6ff16410ac7dbd26190e0fdad3be8a81617e6d3683653a874372239b7c33910.scope\": RecentStats: unable to find data in memory cache]" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.373194 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.504452 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-combined-ca-bundle\") pod \"f40981fe-a91e-45ef-a39c-ef449c19c751\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.504810 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-config-data\") pod \"f40981fe-a91e-45ef-a39c-ef449c19c751\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.504909 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z7k9\" (UniqueName: \"kubernetes.io/projected/f40981fe-a91e-45ef-a39c-ef449c19c751-kube-api-access-8z7k9\") pod \"f40981fe-a91e-45ef-a39c-ef449c19c751\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.504998 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-log-httpd\") pod \"f40981fe-a91e-45ef-a39c-ef449c19c751\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.505052 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-sg-core-conf-yaml\") pod \"f40981fe-a91e-45ef-a39c-ef449c19c751\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.505082 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-scripts\") pod \"f40981fe-a91e-45ef-a39c-ef449c19c751\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.505318 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-run-httpd\") pod \"f40981fe-a91e-45ef-a39c-ef449c19c751\" (UID: \"f40981fe-a91e-45ef-a39c-ef449c19c751\") " Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.505848 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f40981fe-a91e-45ef-a39c-ef449c19c751" (UID: "f40981fe-a91e-45ef-a39c-ef449c19c751"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.507324 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f40981fe-a91e-45ef-a39c-ef449c19c751" (UID: "f40981fe-a91e-45ef-a39c-ef449c19c751"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.525821 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f40981fe-a91e-45ef-a39c-ef449c19c751-kube-api-access-8z7k9" (OuterVolumeSpecName: "kube-api-access-8z7k9") pod "f40981fe-a91e-45ef-a39c-ef449c19c751" (UID: "f40981fe-a91e-45ef-a39c-ef449c19c751"). InnerVolumeSpecName "kube-api-access-8z7k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.530429 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-scripts" (OuterVolumeSpecName: "scripts") pod "f40981fe-a91e-45ef-a39c-ef449c19c751" (UID: "f40981fe-a91e-45ef-a39c-ef449c19c751"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.568801 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f40981fe-a91e-45ef-a39c-ef449c19c751" (UID: "f40981fe-a91e-45ef-a39c-ef449c19c751"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.608227 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z7k9\" (UniqueName: \"kubernetes.io/projected/f40981fe-a91e-45ef-a39c-ef449c19c751-kube-api-access-8z7k9\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.608268 4954 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.608281 4954 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.608291 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.608303 4954 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f40981fe-a91e-45ef-a39c-ef449c19c751-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.635859 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f40981fe-a91e-45ef-a39c-ef449c19c751" (UID: "f40981fe-a91e-45ef-a39c-ef449c19c751"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.677505 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-config-data" (OuterVolumeSpecName: "config-data") pod "f40981fe-a91e-45ef-a39c-ef449c19c751" (UID: "f40981fe-a91e-45ef-a39c-ef449c19c751"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.710823 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.711169 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40981fe-a91e-45ef-a39c-ef449c19c751-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.974395 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f40981fe-a91e-45ef-a39c-ef449c19c751","Type":"ContainerDied","Data":"c40b4d4ffab2164339fd9b39a0f189ae1c24e0ae7e498e26304d31503c1866f6"} Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.974471 4954 scope.go:117] "RemoveContainer" containerID="109b477bde6f46241718782f183745f21cb35376ea49165b7bd443649b0f5566" Dec 09 17:29:11 crc kubenswrapper[4954]: I1209 17:29:11.974738 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.018337 4954 scope.go:117] "RemoveContainer" containerID="0a27558d4d698f6c77257057c2215c9ffcd635d5dec3a5dec26aef26c8a9d5ec" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.035733 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.047014 4954 scope.go:117] "RemoveContainer" containerID="d39c771e3ee46639b0fd9a5c1a37b4b6219ca751e5bca6c01b8914c849f52c61" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.055020 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.073666 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074240 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" containerName="init" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074258 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" containerName="init" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074270 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="ceilometer-central-agent" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074277 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="ceilometer-central-agent" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074300 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="ceilometer-notification-agent" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074307 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="ceilometer-notification-agent" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074315 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="sg-core" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074321 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="sg-core" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074337 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" containerName="dnsmasq-dns" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074343 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" containerName="dnsmasq-dns" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074350 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b8a6386-1d71-4193-8d87-303561f4b34a" containerName="heat-api" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074356 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b8a6386-1d71-4193-8d87-303561f4b34a" containerName="heat-api" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074365 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" containerName="heat-cfnapi" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074371 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" containerName="heat-cfnapi" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074383 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91fe1cc4-243c-473f-a86a-2a961648027f" containerName="heat-cfnapi" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074388 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="91fe1cc4-243c-473f-a86a-2a961648027f" containerName="heat-cfnapi" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074407 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5678723-dfca-41a0-ac7b-3a4418060e27" containerName="heat-api" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074413 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5678723-dfca-41a0-ac7b-3a4418060e27" containerName="heat-api" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074422 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" containerName="heat-cfnapi" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074428 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" containerName="heat-cfnapi" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074436 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="proxy-httpd" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074442 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="proxy-httpd" Dec 09 17:29:12 crc kubenswrapper[4954]: E1209 17:29:12.074466 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5678723-dfca-41a0-ac7b-3a4418060e27" containerName="heat-api" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074472 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5678723-dfca-41a0-ac7b-3a4418060e27" containerName="heat-api" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074705 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5678723-dfca-41a0-ac7b-3a4418060e27" containerName="heat-api" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074722 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="sg-core" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074734 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="ceilometer-notification-agent" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074743 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="proxy-httpd" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074754 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" containerName="heat-cfnapi" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074777 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d002f0cb-84e5-4bd3-a5f9-4ac71789e0be" containerName="dnsmasq-dns" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074787 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="91fe1cc4-243c-473f-a86a-2a961648027f" containerName="heat-cfnapi" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074801 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" containerName="ceilometer-central-agent" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074809 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b8a6386-1d71-4193-8d87-303561f4b34a" containerName="heat-api" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.074821 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae072c84-04f0-48a3-a9ee-1068a9ae7a34" containerName="heat-cfnapi" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.075257 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5678723-dfca-41a0-ac7b-3a4418060e27" containerName="heat-api" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.076896 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.081534 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.081785 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.085703 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.107673 4954 scope.go:117] "RemoveContainer" containerID="da7e20528413ebb59131126e1769f5a793003b872b669e19c22f86bc40727556" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.145559 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.145917 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l54l4\" (UniqueName: \"kubernetes.io/projected/77cd40ac-9ed6-4b59-a223-1fca57aea614-kube-api-access-l54l4\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.146044 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-run-httpd\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.146151 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.146270 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-config-data\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.146367 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-scripts\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.146480 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-log-httpd\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.175960 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f40981fe-a91e-45ef-a39c-ef449c19c751" path="/var/lib/kubelet/pods/f40981fe-a91e-45ef-a39c-ef449c19c751/volumes" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.248731 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l54l4\" (UniqueName: \"kubernetes.io/projected/77cd40ac-9ed6-4b59-a223-1fca57aea614-kube-api-access-l54l4\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.249163 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-run-httpd\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.249206 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.249268 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-config-data\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.249290 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-scripts\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.249335 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-log-httpd\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.249393 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.249838 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-run-httpd\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.251628 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-log-httpd\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.255806 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.262785 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.263523 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-scripts\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.269748 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l54l4\" (UniqueName: \"kubernetes.io/projected/77cd40ac-9ed6-4b59-a223-1fca57aea614-kube-api-access-l54l4\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.269991 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-config-data\") pod \"ceilometer-0\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " pod="openstack/ceilometer-0" Dec 09 17:29:12 crc kubenswrapper[4954]: I1209 17:29:12.432901 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:13 crc kubenswrapper[4954]: I1209 17:29:13.029500 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:13 crc kubenswrapper[4954]: I1209 17:29:13.999646 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77cd40ac-9ed6-4b59-a223-1fca57aea614","Type":"ContainerStarted","Data":"eae349563d3ed2796dd1de7e0f192222afcf718e6f7de8a8bb1cc1f4c247689a"} Dec 09 17:29:13 crc kubenswrapper[4954]: I1209 17:29:13.999969 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77cd40ac-9ed6-4b59-a223-1fca57aea614","Type":"ContainerStarted","Data":"6e45d47832d353ab15a71bb8018ee2326c96e25dd168e6baeb296943db6b1f73"} Dec 09 17:29:14 crc kubenswrapper[4954]: I1209 17:29:14.000999 4954 generic.go:334] "Generic (PLEG): container finished" podID="24d42cd5-dc37-4675-87b9-91dcb7391eed" containerID="289fc5c5a373dc0fbb22475c3925fb11dc75b13b2f17fb7b28cdf313f453841a" exitCode=0 Dec 09 17:29:14 crc kubenswrapper[4954]: I1209 17:29:14.001028 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ln5zl" event={"ID":"24d42cd5-dc37-4675-87b9-91dcb7391eed","Type":"ContainerDied","Data":"289fc5c5a373dc0fbb22475c3925fb11dc75b13b2f17fb7b28cdf313f453841a"} Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.546132 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.640740 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-combined-ca-bundle\") pod \"24d42cd5-dc37-4675-87b9-91dcb7391eed\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.640824 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-config-data\") pod \"24d42cd5-dc37-4675-87b9-91dcb7391eed\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.640958 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78fx2\" (UniqueName: \"kubernetes.io/projected/24d42cd5-dc37-4675-87b9-91dcb7391eed-kube-api-access-78fx2\") pod \"24d42cd5-dc37-4675-87b9-91dcb7391eed\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.641059 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-scripts\") pod \"24d42cd5-dc37-4675-87b9-91dcb7391eed\" (UID: \"24d42cd5-dc37-4675-87b9-91dcb7391eed\") " Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.646751 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d42cd5-dc37-4675-87b9-91dcb7391eed-kube-api-access-78fx2" (OuterVolumeSpecName: "kube-api-access-78fx2") pod "24d42cd5-dc37-4675-87b9-91dcb7391eed" (UID: "24d42cd5-dc37-4675-87b9-91dcb7391eed"). InnerVolumeSpecName "kube-api-access-78fx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.650036 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-scripts" (OuterVolumeSpecName: "scripts") pod "24d42cd5-dc37-4675-87b9-91dcb7391eed" (UID: "24d42cd5-dc37-4675-87b9-91dcb7391eed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.676413 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-config-data" (OuterVolumeSpecName: "config-data") pod "24d42cd5-dc37-4675-87b9-91dcb7391eed" (UID: "24d42cd5-dc37-4675-87b9-91dcb7391eed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.679059 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24d42cd5-dc37-4675-87b9-91dcb7391eed" (UID: "24d42cd5-dc37-4675-87b9-91dcb7391eed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.744196 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78fx2\" (UniqueName: \"kubernetes.io/projected/24d42cd5-dc37-4675-87b9-91dcb7391eed-kube-api-access-78fx2\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.744235 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.744247 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:15 crc kubenswrapper[4954]: I1209 17:29:15.744256 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d42cd5-dc37-4675-87b9-91dcb7391eed-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.035972 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ln5zl" event={"ID":"24d42cd5-dc37-4675-87b9-91dcb7391eed","Type":"ContainerDied","Data":"3bb6185f96517e853a1f4c4b427d57d83335534a30dde20f28a396fe8ddbe363"} Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.036049 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bb6185f96517e853a1f4c4b427d57d83335534a30dde20f28a396fe8ddbe363" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.036018 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ln5zl" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.038578 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77cd40ac-9ed6-4b59-a223-1fca57aea614","Type":"ContainerStarted","Data":"01b549bbf87b9ae12200c15ede235ab40964c68816aa64cb29c54f529a01467e"} Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.189744 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 17:29:16 crc kubenswrapper[4954]: E1209 17:29:16.190620 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d42cd5-dc37-4675-87b9-91dcb7391eed" containerName="nova-cell0-conductor-db-sync" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.190639 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d42cd5-dc37-4675-87b9-91dcb7391eed" containerName="nova-cell0-conductor-db-sync" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.190908 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d42cd5-dc37-4675-87b9-91dcb7391eed" containerName="nova-cell0-conductor-db-sync" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.191821 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.194295 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-zj68k" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.194667 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.208172 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.255812 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3b535e-6a1e-4ff0-abdf-be624309bb19-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5f3b535e-6a1e-4ff0-abdf-be624309bb19\") " pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.255957 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4sxv\" (UniqueName: \"kubernetes.io/projected/5f3b535e-6a1e-4ff0-abdf-be624309bb19-kube-api-access-q4sxv\") pod \"nova-cell0-conductor-0\" (UID: \"5f3b535e-6a1e-4ff0-abdf-be624309bb19\") " pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.256035 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3b535e-6a1e-4ff0-abdf-be624309bb19-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5f3b535e-6a1e-4ff0-abdf-be624309bb19\") " pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.358634 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4sxv\" (UniqueName: \"kubernetes.io/projected/5f3b535e-6a1e-4ff0-abdf-be624309bb19-kube-api-access-q4sxv\") pod \"nova-cell0-conductor-0\" (UID: \"5f3b535e-6a1e-4ff0-abdf-be624309bb19\") " pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.359248 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3b535e-6a1e-4ff0-abdf-be624309bb19-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5f3b535e-6a1e-4ff0-abdf-be624309bb19\") " pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.359531 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3b535e-6a1e-4ff0-abdf-be624309bb19-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5f3b535e-6a1e-4ff0-abdf-be624309bb19\") " pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.364681 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3b535e-6a1e-4ff0-abdf-be624309bb19-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5f3b535e-6a1e-4ff0-abdf-be624309bb19\") " pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.374474 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3b535e-6a1e-4ff0-abdf-be624309bb19-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5f3b535e-6a1e-4ff0-abdf-be624309bb19\") " pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.377884 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4sxv\" (UniqueName: \"kubernetes.io/projected/5f3b535e-6a1e-4ff0-abdf-be624309bb19-kube-api-access-q4sxv\") pod \"nova-cell0-conductor-0\" (UID: \"5f3b535e-6a1e-4ff0-abdf-be624309bb19\") " pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:16 crc kubenswrapper[4954]: I1209 17:29:16.512100 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.056499 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77cd40ac-9ed6-4b59-a223-1fca57aea614","Type":"ContainerStarted","Data":"e6b0582a90711f6919a6ae2438498d40e06bc9c1e08e18d5cf406d2df2fcbfe9"} Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.099619 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.426721 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.492376 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msctm\" (UniqueName: \"kubernetes.io/projected/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-kube-api-access-msctm\") pod \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.493121 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-combined-ca-bundle\") pod \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.493217 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data-custom\") pod \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.493304 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data\") pod \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\" (UID: \"cfe2f3db-ef77-459d-8ee8-e0df346b5beb\") " Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.506177 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-kube-api-access-msctm" (OuterVolumeSpecName: "kube-api-access-msctm") pod "cfe2f3db-ef77-459d-8ee8-e0df346b5beb" (UID: "cfe2f3db-ef77-459d-8ee8-e0df346b5beb"). InnerVolumeSpecName "kube-api-access-msctm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.508090 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cfe2f3db-ef77-459d-8ee8-e0df346b5beb" (UID: "cfe2f3db-ef77-459d-8ee8-e0df346b5beb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.510717 4954 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.510745 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msctm\" (UniqueName: \"kubernetes.io/projected/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-kube-api-access-msctm\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.582313 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfe2f3db-ef77-459d-8ee8-e0df346b5beb" (UID: "cfe2f3db-ef77-459d-8ee8-e0df346b5beb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.616063 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.637771 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data" (OuterVolumeSpecName: "config-data") pod "cfe2f3db-ef77-459d-8ee8-e0df346b5beb" (UID: "cfe2f3db-ef77-459d-8ee8-e0df346b5beb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:17 crc kubenswrapper[4954]: I1209 17:29:17.729184 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe2f3db-ef77-459d-8ee8-e0df346b5beb-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.068135 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5f3b535e-6a1e-4ff0-abdf-be624309bb19","Type":"ContainerStarted","Data":"f73c8e3b377fd807a03bcd06afe531999643adac458cf2df5fe45d73e58f130f"} Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.068474 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5f3b535e-6a1e-4ff0-abdf-be624309bb19","Type":"ContainerStarted","Data":"b65922ef5f7816e2c3fd2c20def0bafd2ece5f94b12fa9f57ca13e260d731fbb"} Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.068517 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.071853 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77cd40ac-9ed6-4b59-a223-1fca57aea614","Type":"ContainerStarted","Data":"19e813f54f7b87a3a3c8e4f66fc91b112bd4b2fa46008d0827d83651a82014c5"} Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.072803 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.074507 4954 generic.go:334] "Generic (PLEG): container finished" podID="cfe2f3db-ef77-459d-8ee8-e0df346b5beb" containerID="63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e" exitCode=0 Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.074540 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7b79757c4c-zrc6v" event={"ID":"cfe2f3db-ef77-459d-8ee8-e0df346b5beb","Type":"ContainerDied","Data":"63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e"} Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.074559 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7b79757c4c-zrc6v" event={"ID":"cfe2f3db-ef77-459d-8ee8-e0df346b5beb","Type":"ContainerDied","Data":"78ca2073596ce3e4710dfa2bd77ca138fff6555838919904681917c5209bce6f"} Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.074576 4954 scope.go:117] "RemoveContainer" containerID="63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e" Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.074711 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7b79757c4c-zrc6v" Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.090108 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.090086226 podStartE2EDuration="2.090086226s" podCreationTimestamp="2025-12-09 17:29:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:29:18.085423089 +0000 UTC m=+1954.473596939" watchObservedRunningTime="2025-12-09 17:29:18.090086226 +0000 UTC m=+1954.478260036" Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.123659 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:29:18 crc kubenswrapper[4954]: E1209 17:29:18.123951 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.124882 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.694358755 podStartE2EDuration="6.124844404s" podCreationTimestamp="2025-12-09 17:29:12 +0000 UTC" firstStartedPulling="2025-12-09 17:29:13.030207964 +0000 UTC m=+1949.418381784" lastFinishedPulling="2025-12-09 17:29:17.460693613 +0000 UTC m=+1953.848867433" observedRunningTime="2025-12-09 17:29:18.107866232 +0000 UTC m=+1954.496040052" watchObservedRunningTime="2025-12-09 17:29:18.124844404 +0000 UTC m=+1954.513018234" Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.125317 4954 scope.go:117] "RemoveContainer" containerID="63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e" Dec 09 17:29:18 crc kubenswrapper[4954]: E1209 17:29:18.126156 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e\": container with ID starting with 63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e not found: ID does not exist" containerID="63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e" Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.126194 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e"} err="failed to get container status \"63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e\": rpc error: code = NotFound desc = could not find container \"63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e\": container with ID starting with 63da4a4cb479ebd9ce219e9a0ae89627512764be2006f1623b1a1c2bde00551e not found: ID does not exist" Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.202034 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7b79757c4c-zrc6v"] Dec 09 17:29:18 crc kubenswrapper[4954]: I1209 17:29:18.217491 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-7b79757c4c-zrc6v"] Dec 09 17:29:20 crc kubenswrapper[4954]: I1209 17:29:20.134532 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfe2f3db-ef77-459d-8ee8-e0df346b5beb" path="/var/lib/kubelet/pods/cfe2f3db-ef77-459d-8ee8-e0df346b5beb/volumes" Dec 09 17:29:24 crc kubenswrapper[4954]: I1209 17:29:24.158466 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:24 crc kubenswrapper[4954]: I1209 17:29:24.159636 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="ceilometer-central-agent" containerID="cri-o://eae349563d3ed2796dd1de7e0f192222afcf718e6f7de8a8bb1cc1f4c247689a" gracePeriod=30 Dec 09 17:29:24 crc kubenswrapper[4954]: I1209 17:29:24.160074 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="proxy-httpd" containerID="cri-o://19e813f54f7b87a3a3c8e4f66fc91b112bd4b2fa46008d0827d83651a82014c5" gracePeriod=30 Dec 09 17:29:24 crc kubenswrapper[4954]: I1209 17:29:24.160087 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="ceilometer-notification-agent" containerID="cri-o://01b549bbf87b9ae12200c15ede235ab40964c68816aa64cb29c54f529a01467e" gracePeriod=30 Dec 09 17:29:24 crc kubenswrapper[4954]: I1209 17:29:24.160073 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="sg-core" containerID="cri-o://e6b0582a90711f6919a6ae2438498d40e06bc9c1e08e18d5cf406d2df2fcbfe9" gracePeriod=30 Dec 09 17:29:25 crc kubenswrapper[4954]: I1209 17:29:25.157578 4954 generic.go:334] "Generic (PLEG): container finished" podID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerID="19e813f54f7b87a3a3c8e4f66fc91b112bd4b2fa46008d0827d83651a82014c5" exitCode=0 Dec 09 17:29:25 crc kubenswrapper[4954]: I1209 17:29:25.159239 4954 generic.go:334] "Generic (PLEG): container finished" podID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerID="e6b0582a90711f6919a6ae2438498d40e06bc9c1e08e18d5cf406d2df2fcbfe9" exitCode=2 Dec 09 17:29:25 crc kubenswrapper[4954]: I1209 17:29:25.160245 4954 generic.go:334] "Generic (PLEG): container finished" podID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerID="01b549bbf87b9ae12200c15ede235ab40964c68816aa64cb29c54f529a01467e" exitCode=0 Dec 09 17:29:25 crc kubenswrapper[4954]: I1209 17:29:25.157644 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77cd40ac-9ed6-4b59-a223-1fca57aea614","Type":"ContainerDied","Data":"19e813f54f7b87a3a3c8e4f66fc91b112bd4b2fa46008d0827d83651a82014c5"} Dec 09 17:29:25 crc kubenswrapper[4954]: I1209 17:29:25.160323 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77cd40ac-9ed6-4b59-a223-1fca57aea614","Type":"ContainerDied","Data":"e6b0582a90711f6919a6ae2438498d40e06bc9c1e08e18d5cf406d2df2fcbfe9"} Dec 09 17:29:25 crc kubenswrapper[4954]: I1209 17:29:25.160343 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77cd40ac-9ed6-4b59-a223-1fca57aea614","Type":"ContainerDied","Data":"01b549bbf87b9ae12200c15ede235ab40964c68816aa64cb29c54f529a01467e"} Dec 09 17:29:26 crc kubenswrapper[4954]: I1209 17:29:26.553362 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.027651 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-blmj4"] Dec 09 17:29:27 crc kubenswrapper[4954]: E1209 17:29:27.028522 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe2f3db-ef77-459d-8ee8-e0df346b5beb" containerName="heat-engine" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.028546 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe2f3db-ef77-459d-8ee8-e0df346b5beb" containerName="heat-engine" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.028807 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe2f3db-ef77-459d-8ee8-e0df346b5beb" containerName="heat-engine" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.029851 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.033313 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.033392 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.042957 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-blmj4"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.091534 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2cnj\" (UniqueName: \"kubernetes.io/projected/0b263b3a-d693-45c2-8d18-ff0170ee5851-kube-api-access-g2cnj\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.091764 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-config-data\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.091964 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.092607 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-scripts\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.194254 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2cnj\" (UniqueName: \"kubernetes.io/projected/0b263b3a-d693-45c2-8d18-ff0170ee5851-kube-api-access-g2cnj\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.194308 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-config-data\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.194340 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.194424 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-scripts\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.201374 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-scripts\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.213296 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-config-data\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.229520 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.239954 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2cnj\" (UniqueName: \"kubernetes.io/projected/0b263b3a-d693-45c2-8d18-ff0170ee5851-kube-api-access-g2cnj\") pod \"nova-cell0-cell-mapping-blmj4\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.278830 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.281121 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.285664 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.295694 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.355439 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.411508 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.411981 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036dacf0-3f42-40cd-82d7-1b037268d95b-logs\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.412029 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t42j6\" (UniqueName: \"kubernetes.io/projected/036dacf0-3f42-40cd-82d7-1b037268d95b-kube-api-access-t42j6\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.412173 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-config-data\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.412358 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-6jg6z"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.430643 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6jg6z" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.460365 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-6jg6z"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.494992 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.497060 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.505500 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.511604 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.513300 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.519996 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036dacf0-3f42-40cd-82d7-1b037268d95b-logs\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.520037 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t42j6\" (UniqueName: \"kubernetes.io/projected/036dacf0-3f42-40cd-82d7-1b037268d95b-kube-api-access-t42j6\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.520123 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-config-data\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.520195 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.520233 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/216180c8-daea-4b80-8d94-659e92904a12-operator-scripts\") pod \"aodh-db-create-6jg6z\" (UID: \"216180c8-daea-4b80-8d94-659e92904a12\") " pod="openstack/aodh-db-create-6jg6z" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.520256 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h78bg\" (UniqueName: \"kubernetes.io/projected/216180c8-daea-4b80-8d94-659e92904a12-kube-api-access-h78bg\") pod \"aodh-db-create-6jg6z\" (UID: \"216180c8-daea-4b80-8d94-659e92904a12\") " pod="openstack/aodh-db-create-6jg6z" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.529787 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-config-data\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.530066 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036dacf0-3f42-40cd-82d7-1b037268d95b-logs\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.531615 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.554021 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.558640 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.596967 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t42j6\" (UniqueName: \"kubernetes.io/projected/036dacf0-3f42-40cd-82d7-1b037268d95b-kube-api-access-t42j6\") pod \"nova-api-0\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.622076 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.622151 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-config-data\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.622183 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/216180c8-daea-4b80-8d94-659e92904a12-operator-scripts\") pod \"aodh-db-create-6jg6z\" (UID: \"216180c8-daea-4b80-8d94-659e92904a12\") " pod="openstack/aodh-db-create-6jg6z" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.622201 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3ef8419-65d4-476b-a8b2-b748c502ac6a-logs\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.622222 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h78bg\" (UniqueName: \"kubernetes.io/projected/216180c8-daea-4b80-8d94-659e92904a12-kube-api-access-h78bg\") pod \"aodh-db-create-6jg6z\" (UID: \"216180c8-daea-4b80-8d94-659e92904a12\") " pod="openstack/aodh-db-create-6jg6z" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.622286 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.622382 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.622402 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwf2g\" (UniqueName: \"kubernetes.io/projected/b3ef8419-65d4-476b-a8b2-b748c502ac6a-kube-api-access-wwf2g\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.622427 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnddt\" (UniqueName: \"kubernetes.io/projected/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-kube-api-access-nnddt\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.623361 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/216180c8-daea-4b80-8d94-659e92904a12-operator-scripts\") pod \"aodh-db-create-6jg6z\" (UID: \"216180c8-daea-4b80-8d94-659e92904a12\") " pod="openstack/aodh-db-create-6jg6z" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.647352 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.656706 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h78bg\" (UniqueName: \"kubernetes.io/projected/216180c8-daea-4b80-8d94-659e92904a12-kube-api-access-h78bg\") pod \"aodh-db-create-6jg6z\" (UID: \"216180c8-daea-4b80-8d94-659e92904a12\") " pod="openstack/aodh-db-create-6jg6z" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.689690 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.724865 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.725161 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwf2g\" (UniqueName: \"kubernetes.io/projected/b3ef8419-65d4-476b-a8b2-b748c502ac6a-kube-api-access-wwf2g\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.725235 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnddt\" (UniqueName: \"kubernetes.io/projected/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-kube-api-access-nnddt\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.725315 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.725377 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-config-data\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.725418 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3ef8419-65d4-476b-a8b2-b748c502ac6a-logs\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.725544 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.731975 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3ef8419-65d4-476b-a8b2-b748c502ac6a-logs\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.737145 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.739483 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.746453 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.763502 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-52c6-account-create-update-gcs5d"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.765084 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-52c6-account-create-update-gcs5d" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.769735 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnddt\" (UniqueName: \"kubernetes.io/projected/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-kube-api-access-nnddt\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.771148 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-config-data\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.777120 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.783081 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-52c6-account-create-update-gcs5d"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.791478 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwf2g\" (UniqueName: \"kubernetes.io/projected/b3ef8419-65d4-476b-a8b2-b748c502ac6a-kube-api-access-wwf2g\") pod \"nova-metadata-0\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " pod="openstack/nova-metadata-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.825365 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.826934 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.832426 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/436a9623-3c31-40be-a3bc-e4e3364393ac-operator-scripts\") pod \"aodh-52c6-account-create-update-gcs5d\" (UID: \"436a9623-3c31-40be-a3bc-e4e3364393ac\") " pod="openstack/aodh-52c6-account-create-update-gcs5d" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.832537 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rkzm\" (UniqueName: \"kubernetes.io/projected/436a9623-3c31-40be-a3bc-e4e3364393ac-kube-api-access-4rkzm\") pod \"aodh-52c6-account-create-update-gcs5d\" (UID: \"436a9623-3c31-40be-a3bc-e4e3364393ac\") " pod="openstack/aodh-52c6-account-create-update-gcs5d" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.834772 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.874281 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.928039 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6jg6z" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.937733 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gbvxf"] Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.945993 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpgjn\" (UniqueName: \"kubernetes.io/projected/a92fd537-d2cb-4d18-843e-82a6671896ed-kube-api-access-bpgjn\") pod \"nova-scheduler-0\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.946086 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rkzm\" (UniqueName: \"kubernetes.io/projected/436a9623-3c31-40be-a3bc-e4e3364393ac-kube-api-access-4rkzm\") pod \"aodh-52c6-account-create-update-gcs5d\" (UID: \"436a9623-3c31-40be-a3bc-e4e3364393ac\") " pod="openstack/aodh-52c6-account-create-update-gcs5d" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.946267 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-config-data\") pod \"nova-scheduler-0\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.946720 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.947090 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/436a9623-3c31-40be-a3bc-e4e3364393ac-operator-scripts\") pod \"aodh-52c6-account-create-update-gcs5d\" (UID: \"436a9623-3c31-40be-a3bc-e4e3364393ac\") " pod="openstack/aodh-52c6-account-create-update-gcs5d" Dec 09 17:29:27 crc kubenswrapper[4954]: I1209 17:29:27.948385 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/436a9623-3c31-40be-a3bc-e4e3364393ac-operator-scripts\") pod \"aodh-52c6-account-create-update-gcs5d\" (UID: \"436a9623-3c31-40be-a3bc-e4e3364393ac\") " pod="openstack/aodh-52c6-account-create-update-gcs5d" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.000072 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.009364 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.043473 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rkzm\" (UniqueName: \"kubernetes.io/projected/436a9623-3c31-40be-a3bc-e4e3364393ac-kube-api-access-4rkzm\") pod \"aodh-52c6-account-create-update-gcs5d\" (UID: \"436a9623-3c31-40be-a3bc-e4e3364393ac\") " pod="openstack/aodh-52c6-account-create-update-gcs5d" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.063279 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gbvxf"] Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.152580 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.189794 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-config-data\") pod \"nova-scheduler-0\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.196050 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-52c6-account-create-update-gcs5d" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.196506 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.196815 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpgjn\" (UniqueName: \"kubernetes.io/projected/a92fd537-d2cb-4d18-843e-82a6671896ed-kube-api-access-bpgjn\") pod \"nova-scheduler-0\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.227984 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-config-data\") pod \"nova-scheduler-0\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.247270 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.249023 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpgjn\" (UniqueName: \"kubernetes.io/projected/a92fd537-d2cb-4d18-843e-82a6671896ed-kube-api-access-bpgjn\") pod \"nova-scheduler-0\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.308646 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-config\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.308702 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.308774 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.308812 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm9gw\" (UniqueName: \"kubernetes.io/projected/8198eada-d79f-4b68-877c-35325473c747-kube-api-access-nm9gw\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.308840 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-svc\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.308861 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.413302 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-config\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.413725 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.413822 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.413864 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm9gw\" (UniqueName: \"kubernetes.io/projected/8198eada-d79f-4b68-877c-35325473c747-kube-api-access-nm9gw\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.413929 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-svc\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.413947 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.416192 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-config\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.416551 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.417769 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.418558 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-svc\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.418929 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.469307 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm9gw\" (UniqueName: \"kubernetes.io/projected/8198eada-d79f-4b68-877c-35325473c747-kube-api-access-nm9gw\") pod \"dnsmasq-dns-9b86998b5-gbvxf\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.528385 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.711910 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-blmj4"] Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.719267 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:28 crc kubenswrapper[4954]: W1209 17:29:28.754670 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b263b3a_d693_45c2_8d18_ff0170ee5851.slice/crio-5227668b167088cef44f7f81d6e0daf22848107991456d66c4067935bce953e3 WatchSource:0}: Error finding container 5227668b167088cef44f7f81d6e0daf22848107991456d66c4067935bce953e3: Status 404 returned error can't find the container with id 5227668b167088cef44f7f81d6e0daf22848107991456d66c4067935bce953e3 Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.833857 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:29:28 crc kubenswrapper[4954]: I1209 17:29:28.965765 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-6jg6z"] Dec 09 17:29:29 crc kubenswrapper[4954]: W1209 17:29:29.002632 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod216180c8_daea_4b80_8d94_659e92904a12.slice/crio-e53d3c33084498a5c7f8134833db8eb262b7f6240db3bac96f07cc9a831b3a9c WatchSource:0}: Error finding container e53d3c33084498a5c7f8134833db8eb262b7f6240db3bac96f07cc9a831b3a9c: Status 404 returned error can't find the container with id e53d3c33084498a5c7f8134833db8eb262b7f6240db3bac96f07cc9a831b3a9c Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.134798 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.272025 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036dacf0-3f42-40cd-82d7-1b037268d95b","Type":"ContainerStarted","Data":"8a09c2adcee8e4a39979fb264d2e53bc10f8dac38b0d33c60ebad5f113bd040a"} Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.276007 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.278692 4954 generic.go:334] "Generic (PLEG): container finished" podID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerID="eae349563d3ed2796dd1de7e0f192222afcf718e6f7de8a8bb1cc1f4c247689a" exitCode=0 Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.278788 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77cd40ac-9ed6-4b59-a223-1fca57aea614","Type":"ContainerDied","Data":"eae349563d3ed2796dd1de7e0f192222afcf718e6f7de8a8bb1cc1f4c247689a"} Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.297627 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-blmj4" event={"ID":"0b263b3a-d693-45c2-8d18-ff0170ee5851","Type":"ContainerStarted","Data":"f38e068ac82370a127fa322b93040f4b78057d4ffd9da46a8d332807fbb35d2a"} Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.297697 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-blmj4" event={"ID":"0b263b3a-d693-45c2-8d18-ff0170ee5851","Type":"ContainerStarted","Data":"5227668b167088cef44f7f81d6e0daf22848107991456d66c4067935bce953e3"} Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.314753 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5b4bfb0c-c99c-4c69-bc82-586084c33d2a","Type":"ContainerStarted","Data":"5e900716bf1eb273d5f31dc027ab6f2d3bd45843398caecaa6b7308f2b7b22d8"} Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.325158 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-6jg6z" event={"ID":"216180c8-daea-4b80-8d94-659e92904a12","Type":"ContainerStarted","Data":"e53d3c33084498a5c7f8134833db8eb262b7f6240db3bac96f07cc9a831b3a9c"} Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.339500 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-blmj4" podStartSLOduration=3.339471684 podStartE2EDuration="3.339471684s" podCreationTimestamp="2025-12-09 17:29:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:29:29.326956031 +0000 UTC m=+1965.715129851" watchObservedRunningTime="2025-12-09 17:29:29.339471684 +0000 UTC m=+1965.727645504" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.453849 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-4gb2h"] Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.455703 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.459902 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.459943 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.464346 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-4gb2h"] Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.497834 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.497925 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbvt4\" (UniqueName: \"kubernetes.io/projected/884b0d29-b0dd-48bf-bb02-51cbb4503b41-kube-api-access-nbvt4\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.497953 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-config-data\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.498011 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-scripts\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.605929 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbvt4\" (UniqueName: \"kubernetes.io/projected/884b0d29-b0dd-48bf-bb02-51cbb4503b41-kube-api-access-nbvt4\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.606247 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-config-data\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.606317 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-scripts\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.606555 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.640459 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-scripts\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.640691 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.652070 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbvt4\" (UniqueName: \"kubernetes.io/projected/884b0d29-b0dd-48bf-bb02-51cbb4503b41-kube-api-access-nbvt4\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.658215 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.660988 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-config-data\") pod \"nova-cell1-conductor-db-sync-4gb2h\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.684050 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-52c6-account-create-update-gcs5d"] Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.709352 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gbvxf"] Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.796799 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:29 crc kubenswrapper[4954]: I1209 17:29:29.943011 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.023769 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l54l4\" (UniqueName: \"kubernetes.io/projected/77cd40ac-9ed6-4b59-a223-1fca57aea614-kube-api-access-l54l4\") pod \"77cd40ac-9ed6-4b59-a223-1fca57aea614\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.024681 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-sg-core-conf-yaml\") pod \"77cd40ac-9ed6-4b59-a223-1fca57aea614\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.024764 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-run-httpd\") pod \"77cd40ac-9ed6-4b59-a223-1fca57aea614\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.024819 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-scripts\") pod \"77cd40ac-9ed6-4b59-a223-1fca57aea614\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.024950 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-combined-ca-bundle\") pod \"77cd40ac-9ed6-4b59-a223-1fca57aea614\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.025075 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-config-data\") pod \"77cd40ac-9ed6-4b59-a223-1fca57aea614\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.025106 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-log-httpd\") pod \"77cd40ac-9ed6-4b59-a223-1fca57aea614\" (UID: \"77cd40ac-9ed6-4b59-a223-1fca57aea614\") " Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.028402 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "77cd40ac-9ed6-4b59-a223-1fca57aea614" (UID: "77cd40ac-9ed6-4b59-a223-1fca57aea614"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.030213 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "77cd40ac-9ed6-4b59-a223-1fca57aea614" (UID: "77cd40ac-9ed6-4b59-a223-1fca57aea614"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.040021 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77cd40ac-9ed6-4b59-a223-1fca57aea614-kube-api-access-l54l4" (OuterVolumeSpecName: "kube-api-access-l54l4") pod "77cd40ac-9ed6-4b59-a223-1fca57aea614" (UID: "77cd40ac-9ed6-4b59-a223-1fca57aea614"). InnerVolumeSpecName "kube-api-access-l54l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.052818 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-scripts" (OuterVolumeSpecName: "scripts") pod "77cd40ac-9ed6-4b59-a223-1fca57aea614" (UID: "77cd40ac-9ed6-4b59-a223-1fca57aea614"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.128409 4954 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.129104 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "77cd40ac-9ed6-4b59-a223-1fca57aea614" (UID: "77cd40ac-9ed6-4b59-a223-1fca57aea614"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.129852 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l54l4\" (UniqueName: \"kubernetes.io/projected/77cd40ac-9ed6-4b59-a223-1fca57aea614-kube-api-access-l54l4\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.129881 4954 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77cd40ac-9ed6-4b59-a223-1fca57aea614-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.129925 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.232559 4954 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.348969 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77cd40ac-9ed6-4b59-a223-1fca57aea614" (UID: "77cd40ac-9ed6-4b59-a223-1fca57aea614"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.375020 4954 generic.go:334] "Generic (PLEG): container finished" podID="216180c8-daea-4b80-8d94-659e92904a12" containerID="b399558fc9a93c7bdefe8831a86a84e5442ffe70b650a46d464c11a06da8c3bc" exitCode=0 Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.375133 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-6jg6z" event={"ID":"216180c8-daea-4b80-8d94-659e92904a12","Type":"ContainerDied","Data":"b399558fc9a93c7bdefe8831a86a84e5442ffe70b650a46d464c11a06da8c3bc"} Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.381751 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3ef8419-65d4-476b-a8b2-b748c502ac6a","Type":"ContainerStarted","Data":"303870c1a1919ba1dd71c9555536d6f31275fe1e206006e71a5a3f7cd4e12049"} Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.398204 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77cd40ac-9ed6-4b59-a223-1fca57aea614","Type":"ContainerDied","Data":"6e45d47832d353ab15a71bb8018ee2326c96e25dd168e6baeb296943db6b1f73"} Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.398272 4954 scope.go:117] "RemoveContainer" containerID="19e813f54f7b87a3a3c8e4f66fc91b112bd4b2fa46008d0827d83651a82014c5" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.398470 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.409866 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" event={"ID":"8198eada-d79f-4b68-877c-35325473c747","Type":"ContainerStarted","Data":"4e36f587cf39ff7f930d3ec1e509aebba3b5980055890d95de5eb610dbbe55ee"} Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.439834 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.485378 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a92fd537-d2cb-4d18-843e-82a6671896ed","Type":"ContainerStarted","Data":"8fa72e19ca13c3843d3a670b601bf71ff8652e350e7a64d0696be0415282efd0"} Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.508034 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-52c6-account-create-update-gcs5d" event={"ID":"436a9623-3c31-40be-a3bc-e4e3364393ac","Type":"ContainerStarted","Data":"953aa7fffb40e256fc65efcc4a796694ffb9e0a2029efcdf6bb29eb5e034ce7b"} Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.512410 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-config-data" (OuterVolumeSpecName: "config-data") pod "77cd40ac-9ed6-4b59-a223-1fca57aea614" (UID: "77cd40ac-9ed6-4b59-a223-1fca57aea614"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.543356 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77cd40ac-9ed6-4b59-a223-1fca57aea614-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.594934 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-4gb2h"] Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.739821 4954 scope.go:117] "RemoveContainer" containerID="e6b0582a90711f6919a6ae2438498d40e06bc9c1e08e18d5cf406d2df2fcbfe9" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.858786 4954 scope.go:117] "RemoveContainer" containerID="01b549bbf87b9ae12200c15ede235ab40964c68816aa64cb29c54f529a01467e" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.868986 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.920647 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.940575 4954 scope.go:117] "RemoveContainer" containerID="eae349563d3ed2796dd1de7e0f192222afcf718e6f7de8a8bb1cc1f4c247689a" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.958682 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:30 crc kubenswrapper[4954]: E1209 17:29:30.959368 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="sg-core" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.959395 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="sg-core" Dec 09 17:29:30 crc kubenswrapper[4954]: E1209 17:29:30.959421 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="ceilometer-central-agent" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.959427 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="ceilometer-central-agent" Dec 09 17:29:30 crc kubenswrapper[4954]: E1209 17:29:30.959450 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="proxy-httpd" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.959458 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="proxy-httpd" Dec 09 17:29:30 crc kubenswrapper[4954]: E1209 17:29:30.959499 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="ceilometer-notification-agent" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.959515 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="ceilometer-notification-agent" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.959798 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="proxy-httpd" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.959825 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="ceilometer-central-agent" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.959833 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="ceilometer-notification-agent" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.959845 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" containerName="sg-core" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.962087 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.970101 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.970295 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 17:29:30 crc kubenswrapper[4954]: I1209 17:29:30.989544 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.091439 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-config-data\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.094968 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.095214 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-run-httpd\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.095423 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-log-httpd\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.095707 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-scripts\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.095741 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.095815 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tw7z\" (UniqueName: \"kubernetes.io/projected/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-kube-api-access-7tw7z\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.198002 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.198130 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-run-httpd\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.198214 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-log-httpd\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.198290 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-scripts\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.198314 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.198351 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tw7z\" (UniqueName: \"kubernetes.io/projected/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-kube-api-access-7tw7z\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.198415 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-config-data\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.200238 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-log-httpd\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.200491 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-run-httpd\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.205327 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-config-data\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.208083 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-scripts\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.208553 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.212153 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.223831 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tw7z\" (UniqueName: \"kubernetes.io/projected/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-kube-api-access-7tw7z\") pod \"ceilometer-0\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.296121 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.590080 4954 generic.go:334] "Generic (PLEG): container finished" podID="8198eada-d79f-4b68-877c-35325473c747" containerID="732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454" exitCode=0 Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.590424 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" event={"ID":"8198eada-d79f-4b68-877c-35325473c747","Type":"ContainerDied","Data":"732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454"} Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.620070 4954 generic.go:334] "Generic (PLEG): container finished" podID="436a9623-3c31-40be-a3bc-e4e3364393ac" containerID="c3d8b0f1fea79e6696f51c27a8d5c41ab2f966b35471565ddc1efa2bc808e962" exitCode=0 Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.620375 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-52c6-account-create-update-gcs5d" event={"ID":"436a9623-3c31-40be-a3bc-e4e3364393ac","Type":"ContainerDied","Data":"c3d8b0f1fea79e6696f51c27a8d5c41ab2f966b35471565ddc1efa2bc808e962"} Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.625830 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-4gb2h" event={"ID":"884b0d29-b0dd-48bf-bb02-51cbb4503b41","Type":"ContainerStarted","Data":"8b9de092f02f4523780f2eaf864710b98a587bde05d42ed00dd49e8e9d6a0a48"} Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.625885 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-4gb2h" event={"ID":"884b0d29-b0dd-48bf-bb02-51cbb4503b41","Type":"ContainerStarted","Data":"4a85334dd1861501777416d49af5dc6c6bcdb15594352878b0c618e9563bd8d0"} Dec 09 17:29:31 crc kubenswrapper[4954]: I1209 17:29:31.749235 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-4gb2h" podStartSLOduration=2.749200421 podStartE2EDuration="2.749200421s" podCreationTimestamp="2025-12-09 17:29:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:29:31.690098801 +0000 UTC m=+1968.078272631" watchObservedRunningTime="2025-12-09 17:29:31.749200421 +0000 UTC m=+1968.137374241" Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.055280 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.163090 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77cd40ac-9ed6-4b59-a223-1fca57aea614" path="/var/lib/kubelet/pods/77cd40ac-9ed6-4b59-a223-1fca57aea614/volumes" Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.371777 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.393216 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.455715 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6jg6z" Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.574967 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/216180c8-daea-4b80-8d94-659e92904a12-operator-scripts\") pod \"216180c8-daea-4b80-8d94-659e92904a12\" (UID: \"216180c8-daea-4b80-8d94-659e92904a12\") " Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.575652 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/216180c8-daea-4b80-8d94-659e92904a12-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "216180c8-daea-4b80-8d94-659e92904a12" (UID: "216180c8-daea-4b80-8d94-659e92904a12"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.575699 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h78bg\" (UniqueName: \"kubernetes.io/projected/216180c8-daea-4b80-8d94-659e92904a12-kube-api-access-h78bg\") pod \"216180c8-daea-4b80-8d94-659e92904a12\" (UID: \"216180c8-daea-4b80-8d94-659e92904a12\") " Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.581983 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/216180c8-daea-4b80-8d94-659e92904a12-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.593931 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/216180c8-daea-4b80-8d94-659e92904a12-kube-api-access-h78bg" (OuterVolumeSpecName: "kube-api-access-h78bg") pod "216180c8-daea-4b80-8d94-659e92904a12" (UID: "216180c8-daea-4b80-8d94-659e92904a12"). InnerVolumeSpecName "kube-api-access-h78bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.648909 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f","Type":"ContainerStarted","Data":"c31b450ecd1b3089c6b9e26dfa1b411256a71d50bbc51c43f085199f75def060"} Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.652947 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6jg6z" Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.652975 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-6jg6z" event={"ID":"216180c8-daea-4b80-8d94-659e92904a12","Type":"ContainerDied","Data":"e53d3c33084498a5c7f8134833db8eb262b7f6240db3bac96f07cc9a831b3a9c"} Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.653010 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e53d3c33084498a5c7f8134833db8eb262b7f6240db3bac96f07cc9a831b3a9c" Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.679954 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" event={"ID":"8198eada-d79f-4b68-877c-35325473c747","Type":"ContainerStarted","Data":"bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c"} Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.680006 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.683899 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h78bg\" (UniqueName: \"kubernetes.io/projected/216180c8-daea-4b80-8d94-659e92904a12-kube-api-access-h78bg\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:32 crc kubenswrapper[4954]: I1209 17:29:32.723775 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" podStartSLOduration=5.723571494 podStartE2EDuration="5.723571494s" podCreationTimestamp="2025-12-09 17:29:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:29:32.712587051 +0000 UTC m=+1969.100760871" watchObservedRunningTime="2025-12-09 17:29:32.723571494 +0000 UTC m=+1969.111745314" Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.120522 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:29:33 crc kubenswrapper[4954]: E1209 17:29:33.121123 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.535618 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-52c6-account-create-update-gcs5d" Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.708436 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/436a9623-3c31-40be-a3bc-e4e3364393ac-operator-scripts\") pod \"436a9623-3c31-40be-a3bc-e4e3364393ac\" (UID: \"436a9623-3c31-40be-a3bc-e4e3364393ac\") " Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.708729 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rkzm\" (UniqueName: \"kubernetes.io/projected/436a9623-3c31-40be-a3bc-e4e3364393ac-kube-api-access-4rkzm\") pod \"436a9623-3c31-40be-a3bc-e4e3364393ac\" (UID: \"436a9623-3c31-40be-a3bc-e4e3364393ac\") " Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.709533 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/436a9623-3c31-40be-a3bc-e4e3364393ac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "436a9623-3c31-40be-a3bc-e4e3364393ac" (UID: "436a9623-3c31-40be-a3bc-e4e3364393ac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.711856 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-52c6-account-create-update-gcs5d" Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.712022 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-52c6-account-create-update-gcs5d" event={"ID":"436a9623-3c31-40be-a3bc-e4e3364393ac","Type":"ContainerDied","Data":"953aa7fffb40e256fc65efcc4a796694ffb9e0a2029efcdf6bb29eb5e034ce7b"} Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.712062 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="953aa7fffb40e256fc65efcc4a796694ffb9e0a2029efcdf6bb29eb5e034ce7b" Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.717167 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/436a9623-3c31-40be-a3bc-e4e3364393ac-kube-api-access-4rkzm" (OuterVolumeSpecName: "kube-api-access-4rkzm") pod "436a9623-3c31-40be-a3bc-e4e3364393ac" (UID: "436a9623-3c31-40be-a3bc-e4e3364393ac"). InnerVolumeSpecName "kube-api-access-4rkzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.812669 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rkzm\" (UniqueName: \"kubernetes.io/projected/436a9623-3c31-40be-a3bc-e4e3364393ac-kube-api-access-4rkzm\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:33 crc kubenswrapper[4954]: I1209 17:29:33.812709 4954 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/436a9623-3c31-40be-a3bc-e4e3364393ac-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:35 crc kubenswrapper[4954]: I1209 17:29:35.740638 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036dacf0-3f42-40cd-82d7-1b037268d95b","Type":"ContainerStarted","Data":"b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8"} Dec 09 17:29:35 crc kubenswrapper[4954]: I1209 17:29:35.745279 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3ef8419-65d4-476b-a8b2-b748c502ac6a","Type":"ContainerStarted","Data":"36b1a2121c8e82e96891b1b98f87e77e9c6469ec4338b825764e0ee882f8bba5"} Dec 09 17:29:35 crc kubenswrapper[4954]: I1209 17:29:35.747234 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a92fd537-d2cb-4d18-843e-82a6671896ed","Type":"ContainerStarted","Data":"64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43"} Dec 09 17:29:35 crc kubenswrapper[4954]: I1209 17:29:35.755207 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f","Type":"ContainerStarted","Data":"061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80"} Dec 09 17:29:35 crc kubenswrapper[4954]: I1209 17:29:35.757192 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5b4bfb0c-c99c-4c69-bc82-586084c33d2a","Type":"ContainerStarted","Data":"f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090"} Dec 09 17:29:35 crc kubenswrapper[4954]: I1209 17:29:35.757368 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="5b4bfb0c-c99c-4c69-bc82-586084c33d2a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090" gracePeriod=30 Dec 09 17:29:35 crc kubenswrapper[4954]: I1209 17:29:35.770649 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.217020735 podStartE2EDuration="8.770626583s" podCreationTimestamp="2025-12-09 17:29:27 +0000 UTC" firstStartedPulling="2025-12-09 17:29:29.635131939 +0000 UTC m=+1966.023305759" lastFinishedPulling="2025-12-09 17:29:35.188737787 +0000 UTC m=+1971.576911607" observedRunningTime="2025-12-09 17:29:35.769329033 +0000 UTC m=+1972.157502863" watchObservedRunningTime="2025-12-09 17:29:35.770626583 +0000 UTC m=+1972.158800403" Dec 09 17:29:35 crc kubenswrapper[4954]: I1209 17:29:35.801100 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.765387056 podStartE2EDuration="8.801070016s" podCreationTimestamp="2025-12-09 17:29:27 +0000 UTC" firstStartedPulling="2025-12-09 17:29:29.155242505 +0000 UTC m=+1965.543416325" lastFinishedPulling="2025-12-09 17:29:35.190925465 +0000 UTC m=+1971.579099285" observedRunningTime="2025-12-09 17:29:35.790091803 +0000 UTC m=+1972.178265623" watchObservedRunningTime="2025-12-09 17:29:35.801070016 +0000 UTC m=+1972.189243836" Dec 09 17:29:36 crc kubenswrapper[4954]: I1209 17:29:36.770278 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036dacf0-3f42-40cd-82d7-1b037268d95b","Type":"ContainerStarted","Data":"cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0"} Dec 09 17:29:36 crc kubenswrapper[4954]: I1209 17:29:36.775981 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3ef8419-65d4-476b-a8b2-b748c502ac6a","Type":"ContainerStarted","Data":"2687865020a55b1292f6b82e56a2bf2e350f19d5ff9fa2be1260cf452acb368b"} Dec 09 17:29:36 crc kubenswrapper[4954]: I1209 17:29:36.776219 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" containerName="nova-metadata-log" containerID="cri-o://36b1a2121c8e82e96891b1b98f87e77e9c6469ec4338b825764e0ee882f8bba5" gracePeriod=30 Dec 09 17:29:36 crc kubenswrapper[4954]: I1209 17:29:36.776264 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" containerName="nova-metadata-metadata" containerID="cri-o://2687865020a55b1292f6b82e56a2bf2e350f19d5ff9fa2be1260cf452acb368b" gracePeriod=30 Dec 09 17:29:36 crc kubenswrapper[4954]: I1209 17:29:36.844108 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.006166061 podStartE2EDuration="9.844078639s" podCreationTimestamp="2025-12-09 17:29:27 +0000 UTC" firstStartedPulling="2025-12-09 17:29:29.352930025 +0000 UTC m=+1965.741103845" lastFinishedPulling="2025-12-09 17:29:35.190842603 +0000 UTC m=+1971.579016423" observedRunningTime="2025-12-09 17:29:36.838944628 +0000 UTC m=+1973.227118448" watchObservedRunningTime="2025-12-09 17:29:36.844078639 +0000 UTC m=+1973.232252459" Dec 09 17:29:36 crc kubenswrapper[4954]: I1209 17:29:36.852373 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.521277882 podStartE2EDuration="9.852348288s" podCreationTimestamp="2025-12-09 17:29:27 +0000 UTC" firstStartedPulling="2025-12-09 17:29:28.859847509 +0000 UTC m=+1965.248021329" lastFinishedPulling="2025-12-09 17:29:35.190917915 +0000 UTC m=+1971.579091735" observedRunningTime="2025-12-09 17:29:36.811814068 +0000 UTC m=+1973.199987888" watchObservedRunningTime="2025-12-09 17:29:36.852348288 +0000 UTC m=+1973.240522108" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.651273 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.651753 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.790467 4954 generic.go:334] "Generic (PLEG): container finished" podID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" containerID="2687865020a55b1292f6b82e56a2bf2e350f19d5ff9fa2be1260cf452acb368b" exitCode=0 Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.790505 4954 generic.go:334] "Generic (PLEG): container finished" podID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" containerID="36b1a2121c8e82e96891b1b98f87e77e9c6469ec4338b825764e0ee882f8bba5" exitCode=143 Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.790558 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3ef8419-65d4-476b-a8b2-b748c502ac6a","Type":"ContainerDied","Data":"2687865020a55b1292f6b82e56a2bf2e350f19d5ff9fa2be1260cf452acb368b"} Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.790662 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3ef8419-65d4-476b-a8b2-b748c502ac6a","Type":"ContainerDied","Data":"36b1a2121c8e82e96891b1b98f87e77e9c6469ec4338b825764e0ee882f8bba5"} Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.793196 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f","Type":"ContainerStarted","Data":"1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb"} Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.873262 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-6q2z2"] Dec 09 17:29:37 crc kubenswrapper[4954]: E1209 17:29:37.874124 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436a9623-3c31-40be-a3bc-e4e3364393ac" containerName="mariadb-account-create-update" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.874147 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="436a9623-3c31-40be-a3bc-e4e3364393ac" containerName="mariadb-account-create-update" Dec 09 17:29:37 crc kubenswrapper[4954]: E1209 17:29:37.874162 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="216180c8-daea-4b80-8d94-659e92904a12" containerName="mariadb-database-create" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.874169 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="216180c8-daea-4b80-8d94-659e92904a12" containerName="mariadb-database-create" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.874405 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="436a9623-3c31-40be-a3bc-e4e3364393ac" containerName="mariadb-account-create-update" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.874429 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="216180c8-daea-4b80-8d94-659e92904a12" containerName="mariadb-database-create" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.878432 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.888617 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.888942 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.889150 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.894479 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-6q2z2"] Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.896859 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-hh57v" Dec 09 17:29:37 crc kubenswrapper[4954]: I1209 17:29:37.958715 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.002180 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.026696 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-combined-ca-bundle\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.027020 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7zr8\" (UniqueName: \"kubernetes.io/projected/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-kube-api-access-s7zr8\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.027226 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-scripts\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.027436 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-config-data\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.129183 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwf2g\" (UniqueName: \"kubernetes.io/projected/b3ef8419-65d4-476b-a8b2-b748c502ac6a-kube-api-access-wwf2g\") pod \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.129463 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3ef8419-65d4-476b-a8b2-b748c502ac6a-logs\") pod \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.129587 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-config-data\") pod \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.129658 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-combined-ca-bundle\") pod \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\" (UID: \"b3ef8419-65d4-476b-a8b2-b748c502ac6a\") " Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.130084 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-scripts\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.130128 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-config-data\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.130261 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-combined-ca-bundle\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.130298 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7zr8\" (UniqueName: \"kubernetes.io/projected/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-kube-api-access-s7zr8\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.130421 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3ef8419-65d4-476b-a8b2-b748c502ac6a-logs" (OuterVolumeSpecName: "logs") pod "b3ef8419-65d4-476b-a8b2-b748c502ac6a" (UID: "b3ef8419-65d4-476b-a8b2-b748c502ac6a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.145495 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-combined-ca-bundle\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.147809 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-scripts\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.160886 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3ef8419-65d4-476b-a8b2-b748c502ac6a-kube-api-access-wwf2g" (OuterVolumeSpecName: "kube-api-access-wwf2g") pod "b3ef8419-65d4-476b-a8b2-b748c502ac6a" (UID: "b3ef8419-65d4-476b-a8b2-b748c502ac6a"). InnerVolumeSpecName "kube-api-access-wwf2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.163371 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-config-data\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.184563 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7zr8\" (UniqueName: \"kubernetes.io/projected/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-kube-api-access-s7zr8\") pod \"aodh-db-sync-6q2z2\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.228285 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-config-data" (OuterVolumeSpecName: "config-data") pod "b3ef8419-65d4-476b-a8b2-b748c502ac6a" (UID: "b3ef8419-65d4-476b-a8b2-b748c502ac6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.233586 4954 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3ef8419-65d4-476b-a8b2-b748c502ac6a-logs\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.233857 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.234002 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwf2g\" (UniqueName: \"kubernetes.io/projected/b3ef8419-65d4-476b-a8b2-b748c502ac6a-kube-api-access-wwf2g\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.235801 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3ef8419-65d4-476b-a8b2-b748c502ac6a" (UID: "b3ef8419-65d4-476b-a8b2-b748c502ac6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.279504 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.337233 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ef8419-65d4-476b-a8b2-b748c502ac6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.531154 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.531733 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.594619 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.721909 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.741985 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.221:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.742359 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.221:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.816660 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-9zsk9"] Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.817717 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" podUID="5b762821-719b-451b-9b19-4f1ae1b12994" containerName="dnsmasq-dns" containerID="cri-o://0bbe12244509bb8e93cb6731a09e52e88cadfef4588187c14193213cfbb07b4d" gracePeriod=10 Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.831300 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3ef8419-65d4-476b-a8b2-b748c502ac6a","Type":"ContainerDied","Data":"303870c1a1919ba1dd71c9555536d6f31275fe1e206006e71a5a3f7cd4e12049"} Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.831385 4954 scope.go:117] "RemoveContainer" containerID="2687865020a55b1292f6b82e56a2bf2e350f19d5ff9fa2be1260cf452acb368b" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.831626 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.892354 4954 scope.go:117] "RemoveContainer" containerID="36b1a2121c8e82e96891b1b98f87e77e9c6469ec4338b825764e0ee882f8bba5" Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.907320 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-6q2z2"] Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.926410 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:38 crc kubenswrapper[4954]: I1209 17:29:38.970868 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.002008 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:39 crc kubenswrapper[4954]: E1209 17:29:39.002909 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" containerName="nova-metadata-log" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.002931 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" containerName="nova-metadata-log" Dec 09 17:29:39 crc kubenswrapper[4954]: E1209 17:29:39.002983 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" containerName="nova-metadata-metadata" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.002990 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" containerName="nova-metadata-metadata" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.003478 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" containerName="nova-metadata-log" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.003516 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" containerName="nova-metadata-metadata" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.004942 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.024041 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.024308 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.034875 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.127924 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.214280 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.214437 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.214753 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/575a8ef8-bd25-4e2c-babe-36d73a352084-logs\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.214804 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-config-data\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.214909 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm7gd\" (UniqueName: \"kubernetes.io/projected/575a8ef8-bd25-4e2c-babe-36d73a352084-kube-api-access-zm7gd\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.316940 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/575a8ef8-bd25-4e2c-babe-36d73a352084-logs\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.317276 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-config-data\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.317329 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm7gd\" (UniqueName: \"kubernetes.io/projected/575a8ef8-bd25-4e2c-babe-36d73a352084-kube-api-access-zm7gd\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.317390 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.317443 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.324519 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-config-data\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.324856 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/575a8ef8-bd25-4e2c-babe-36d73a352084-logs\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.331434 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.346367 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.352195 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm7gd\" (UniqueName: \"kubernetes.io/projected/575a8ef8-bd25-4e2c-babe-36d73a352084-kube-api-access-zm7gd\") pod \"nova-metadata-0\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.563325 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.888444 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6q2z2" event={"ID":"f107a10b-0481-41b9-bcfa-7e5bef86ca1a","Type":"ContainerStarted","Data":"af0feae622680ba4f2085c6b0360cb05a412b1d4ee8fa3549bf822472fe12aa2"} Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.957397 4954 generic.go:334] "Generic (PLEG): container finished" podID="5b762821-719b-451b-9b19-4f1ae1b12994" containerID="0bbe12244509bb8e93cb6731a09e52e88cadfef4588187c14193213cfbb07b4d" exitCode=0 Dec 09 17:29:39 crc kubenswrapper[4954]: I1209 17:29:39.958575 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" event={"ID":"5b762821-719b-451b-9b19-4f1ae1b12994","Type":"ContainerDied","Data":"0bbe12244509bb8e93cb6731a09e52e88cadfef4588187c14193213cfbb07b4d"} Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.115895 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.161772 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3ef8419-65d4-476b-a8b2-b748c502ac6a" path="/var/lib/kubelet/pods/b3ef8419-65d4-476b-a8b2-b748c502ac6a/volumes" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.260232 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-svc\") pod \"5b762821-719b-451b-9b19-4f1ae1b12994\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.260314 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-swift-storage-0\") pod \"5b762821-719b-451b-9b19-4f1ae1b12994\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.260378 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-config\") pod \"5b762821-719b-451b-9b19-4f1ae1b12994\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.260398 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9687x\" (UniqueName: \"kubernetes.io/projected/5b762821-719b-451b-9b19-4f1ae1b12994-kube-api-access-9687x\") pod \"5b762821-719b-451b-9b19-4f1ae1b12994\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.260454 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-sb\") pod \"5b762821-719b-451b-9b19-4f1ae1b12994\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.260650 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-nb\") pod \"5b762821-719b-451b-9b19-4f1ae1b12994\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.294108 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b762821-719b-451b-9b19-4f1ae1b12994-kube-api-access-9687x" (OuterVolumeSpecName: "kube-api-access-9687x") pod "5b762821-719b-451b-9b19-4f1ae1b12994" (UID: "5b762821-719b-451b-9b19-4f1ae1b12994"). InnerVolumeSpecName "kube-api-access-9687x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.374225 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5b762821-719b-451b-9b19-4f1ae1b12994" (UID: "5b762821-719b-451b-9b19-4f1ae1b12994"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.377399 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5b762821-719b-451b-9b19-4f1ae1b12994" (UID: "5b762821-719b-451b-9b19-4f1ae1b12994"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.380874 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.382292 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-nb\") pod \"5b762821-719b-451b-9b19-4f1ae1b12994\" (UID: \"5b762821-719b-451b-9b19-4f1ae1b12994\") " Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.383278 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.383303 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9687x\" (UniqueName: \"kubernetes.io/projected/5b762821-719b-451b-9b19-4f1ae1b12994-kube-api-access-9687x\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:40 crc kubenswrapper[4954]: W1209 17:29:40.383411 4954 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5b762821-719b-451b-9b19-4f1ae1b12994/volumes/kubernetes.io~configmap/ovsdbserver-nb Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.383432 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5b762821-719b-451b-9b19-4f1ae1b12994" (UID: "5b762821-719b-451b-9b19-4f1ae1b12994"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.384149 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5b762821-719b-451b-9b19-4f1ae1b12994" (UID: "5b762821-719b-451b-9b19-4f1ae1b12994"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.394336 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5b762821-719b-451b-9b19-4f1ae1b12994" (UID: "5b762821-719b-451b-9b19-4f1ae1b12994"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.407714 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-config" (OuterVolumeSpecName: "config") pod "5b762821-719b-451b-9b19-4f1ae1b12994" (UID: "5b762821-719b-451b-9b19-4f1ae1b12994"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.485460 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.485500 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.485512 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.485522 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b762821-719b-451b-9b19-4f1ae1b12994-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.989005 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" event={"ID":"5b762821-719b-451b-9b19-4f1ae1b12994","Type":"ContainerDied","Data":"f859ca87cc9a98a954010adfdd67557ab77376fdc7a4b8e8c1dd2ceb1d93ae7e"} Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.989037 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-9zsk9" Dec 09 17:29:40 crc kubenswrapper[4954]: I1209 17:29:40.989363 4954 scope.go:117] "RemoveContainer" containerID="0bbe12244509bb8e93cb6731a09e52e88cadfef4588187c14193213cfbb07b4d" Dec 09 17:29:41 crc kubenswrapper[4954]: I1209 17:29:40.998519 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f","Type":"ContainerStarted","Data":"bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824"} Dec 09 17:29:41 crc kubenswrapper[4954]: I1209 17:29:41.011403 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"575a8ef8-bd25-4e2c-babe-36d73a352084","Type":"ContainerStarted","Data":"02dc8eed1072ba3e00495be50a98b0de98d4b6b2ff4e86c340f431447086a557"} Dec 09 17:29:41 crc kubenswrapper[4954]: I1209 17:29:41.011648 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"575a8ef8-bd25-4e2c-babe-36d73a352084","Type":"ContainerStarted","Data":"ee46cd55df0ec6b7af235716aeeeaed62407221624496f16c1c8d605e16317cf"} Dec 09 17:29:41 crc kubenswrapper[4954]: I1209 17:29:41.067685 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-9zsk9"] Dec 09 17:29:41 crc kubenswrapper[4954]: I1209 17:29:41.076141 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-9zsk9"] Dec 09 17:29:41 crc kubenswrapper[4954]: I1209 17:29:41.103499 4954 scope.go:117] "RemoveContainer" containerID="c77240abe49e549301e6ccb00a9526916075b35e9d421271452ee32df1edc332" Dec 09 17:29:42 crc kubenswrapper[4954]: I1209 17:29:42.035366 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"575a8ef8-bd25-4e2c-babe-36d73a352084","Type":"ContainerStarted","Data":"66a361ce9ec5ba09b9840f06c1f868b0602bfa23b58c143e01de3a9f28322b1c"} Dec 09 17:29:42 crc kubenswrapper[4954]: I1209 17:29:42.059956 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.059928743 podStartE2EDuration="4.059928743s" podCreationTimestamp="2025-12-09 17:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:29:42.055068621 +0000 UTC m=+1978.443242441" watchObservedRunningTime="2025-12-09 17:29:42.059928743 +0000 UTC m=+1978.448102563" Dec 09 17:29:42 crc kubenswrapper[4954]: I1209 17:29:42.147633 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b762821-719b-451b-9b19-4f1ae1b12994" path="/var/lib/kubelet/pods/5b762821-719b-451b-9b19-4f1ae1b12994/volumes" Dec 09 17:29:43 crc kubenswrapper[4954]: I1209 17:29:43.057063 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f","Type":"ContainerStarted","Data":"969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8"} Dec 09 17:29:43 crc kubenswrapper[4954]: I1209 17:29:43.057755 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 17:29:43 crc kubenswrapper[4954]: I1209 17:29:43.084874 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.3041142470000002 podStartE2EDuration="13.084848288s" podCreationTimestamp="2025-12-09 17:29:30 +0000 UTC" firstStartedPulling="2025-12-09 17:29:32.198720273 +0000 UTC m=+1968.586894093" lastFinishedPulling="2025-12-09 17:29:41.979454314 +0000 UTC m=+1978.367628134" observedRunningTime="2025-12-09 17:29:43.082218086 +0000 UTC m=+1979.470391906" watchObservedRunningTime="2025-12-09 17:29:43.084848288 +0000 UTC m=+1979.473022108" Dec 09 17:29:44 crc kubenswrapper[4954]: I1209 17:29:44.074637 4954 generic.go:334] "Generic (PLEG): container finished" podID="0b263b3a-d693-45c2-8d18-ff0170ee5851" containerID="f38e068ac82370a127fa322b93040f4b78057d4ffd9da46a8d332807fbb35d2a" exitCode=0 Dec 09 17:29:44 crc kubenswrapper[4954]: I1209 17:29:44.074717 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-blmj4" event={"ID":"0b263b3a-d693-45c2-8d18-ff0170ee5851","Type":"ContainerDied","Data":"f38e068ac82370a127fa322b93040f4b78057d4ffd9da46a8d332807fbb35d2a"} Dec 09 17:29:44 crc kubenswrapper[4954]: I1209 17:29:44.082160 4954 generic.go:334] "Generic (PLEG): container finished" podID="884b0d29-b0dd-48bf-bb02-51cbb4503b41" containerID="8b9de092f02f4523780f2eaf864710b98a587bde05d42ed00dd49e8e9d6a0a48" exitCode=0 Dec 09 17:29:44 crc kubenswrapper[4954]: I1209 17:29:44.082277 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-4gb2h" event={"ID":"884b0d29-b0dd-48bf-bb02-51cbb4503b41","Type":"ContainerDied","Data":"8b9de092f02f4523780f2eaf864710b98a587bde05d42ed00dd49e8e9d6a0a48"} Dec 09 17:29:44 crc kubenswrapper[4954]: I1209 17:29:44.121751 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:29:44 crc kubenswrapper[4954]: I1209 17:29:44.564637 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 17:29:44 crc kubenswrapper[4954]: I1209 17:29:44.565363 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.319976 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.605749 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.635081 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2cnj\" (UniqueName: \"kubernetes.io/projected/0b263b3a-d693-45c2-8d18-ff0170ee5851-kube-api-access-g2cnj\") pod \"0b263b3a-d693-45c2-8d18-ff0170ee5851\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.635772 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-combined-ca-bundle\") pod \"0b263b3a-d693-45c2-8d18-ff0170ee5851\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.635802 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-scripts\") pod \"0b263b3a-d693-45c2-8d18-ff0170ee5851\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.635846 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-config-data\") pod \"0b263b3a-d693-45c2-8d18-ff0170ee5851\" (UID: \"0b263b3a-d693-45c2-8d18-ff0170ee5851\") " Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.657043 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b263b3a-d693-45c2-8d18-ff0170ee5851-kube-api-access-g2cnj" (OuterVolumeSpecName: "kube-api-access-g2cnj") pod "0b263b3a-d693-45c2-8d18-ff0170ee5851" (UID: "0b263b3a-d693-45c2-8d18-ff0170ee5851"). InnerVolumeSpecName "kube-api-access-g2cnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.657451 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-scripts" (OuterVolumeSpecName: "scripts") pod "0b263b3a-d693-45c2-8d18-ff0170ee5851" (UID: "0b263b3a-d693-45c2-8d18-ff0170ee5851"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.708760 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b263b3a-d693-45c2-8d18-ff0170ee5851" (UID: "0b263b3a-d693-45c2-8d18-ff0170ee5851"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.710843 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-config-data" (OuterVolumeSpecName: "config-data") pod "0b263b3a-d693-45c2-8d18-ff0170ee5851" (UID: "0b263b3a-d693-45c2-8d18-ff0170ee5851"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.727836 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.739207 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-combined-ca-bundle\") pod \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.739569 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-scripts\") pod \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.739778 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbvt4\" (UniqueName: \"kubernetes.io/projected/884b0d29-b0dd-48bf-bb02-51cbb4503b41-kube-api-access-nbvt4\") pod \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.739861 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-config-data\") pod \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\" (UID: \"884b0d29-b0dd-48bf-bb02-51cbb4503b41\") " Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.740823 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2cnj\" (UniqueName: \"kubernetes.io/projected/0b263b3a-d693-45c2-8d18-ff0170ee5851-kube-api-access-g2cnj\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.740855 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.740871 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.740884 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b263b3a-d693-45c2-8d18-ff0170ee5851-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.743458 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-scripts" (OuterVolumeSpecName: "scripts") pod "884b0d29-b0dd-48bf-bb02-51cbb4503b41" (UID: "884b0d29-b0dd-48bf-bb02-51cbb4503b41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.744140 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/884b0d29-b0dd-48bf-bb02-51cbb4503b41-kube-api-access-nbvt4" (OuterVolumeSpecName: "kube-api-access-nbvt4") pod "884b0d29-b0dd-48bf-bb02-51cbb4503b41" (UID: "884b0d29-b0dd-48bf-bb02-51cbb4503b41"). InnerVolumeSpecName "kube-api-access-nbvt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.785171 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "884b0d29-b0dd-48bf-bb02-51cbb4503b41" (UID: "884b0d29-b0dd-48bf-bb02-51cbb4503b41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.793941 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-config-data" (OuterVolumeSpecName: "config-data") pod "884b0d29-b0dd-48bf-bb02-51cbb4503b41" (UID: "884b0d29-b0dd-48bf-bb02-51cbb4503b41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.843417 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.843465 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbvt4\" (UniqueName: \"kubernetes.io/projected/884b0d29-b0dd-48bf-bb02-51cbb4503b41-kube-api-access-nbvt4\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.843477 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:45 crc kubenswrapper[4954]: I1209 17:29:45.843498 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/884b0d29-b0dd-48bf-bb02-51cbb4503b41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.118638 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-blmj4" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.119062 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-blmj4" event={"ID":"0b263b3a-d693-45c2-8d18-ff0170ee5851","Type":"ContainerDied","Data":"5227668b167088cef44f7f81d6e0daf22848107991456d66c4067935bce953e3"} Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.125378 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5227668b167088cef44f7f81d6e0daf22848107991456d66c4067935bce953e3" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.157868 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-4gb2h" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.165550 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-4gb2h" event={"ID":"884b0d29-b0dd-48bf-bb02-51cbb4503b41","Type":"ContainerDied","Data":"4a85334dd1861501777416d49af5dc6c6bcdb15594352878b0c618e9563bd8d0"} Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.165626 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a85334dd1861501777416d49af5dc6c6bcdb15594352878b0c618e9563bd8d0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.177034 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6q2z2" event={"ID":"f107a10b-0481-41b9-bcfa-7e5bef86ca1a","Type":"ContainerStarted","Data":"9bd5edc6798b8894974430538ce7f67ab22bb12c2d70de46d5e42f01fe505ae2"} Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.187559 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"7f2aaeb263fa3699e3f6e9d98e811f8ee1865140f8b50100f2a0f8b5ce6f4313"} Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.211005 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-6q2z2" podStartSLOduration=2.829083175 podStartE2EDuration="9.210980443s" podCreationTimestamp="2025-12-09 17:29:37 +0000 UTC" firstStartedPulling="2025-12-09 17:29:38.933660394 +0000 UTC m=+1975.321834214" lastFinishedPulling="2025-12-09 17:29:45.315557662 +0000 UTC m=+1981.703731482" observedRunningTime="2025-12-09 17:29:46.197503992 +0000 UTC m=+1982.585677832" watchObservedRunningTime="2025-12-09 17:29:46.210980443 +0000 UTC m=+1982.599154253" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.233951 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 17:29:46 crc kubenswrapper[4954]: E1209 17:29:46.234881 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b762821-719b-451b-9b19-4f1ae1b12994" containerName="dnsmasq-dns" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.234909 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b762821-719b-451b-9b19-4f1ae1b12994" containerName="dnsmasq-dns" Dec 09 17:29:46 crc kubenswrapper[4954]: E1209 17:29:46.234926 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="884b0d29-b0dd-48bf-bb02-51cbb4503b41" containerName="nova-cell1-conductor-db-sync" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.234937 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="884b0d29-b0dd-48bf-bb02-51cbb4503b41" containerName="nova-cell1-conductor-db-sync" Dec 09 17:29:46 crc kubenswrapper[4954]: E1209 17:29:46.234960 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b263b3a-d693-45c2-8d18-ff0170ee5851" containerName="nova-manage" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.234969 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b263b3a-d693-45c2-8d18-ff0170ee5851" containerName="nova-manage" Dec 09 17:29:46 crc kubenswrapper[4954]: E1209 17:29:46.234997 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b762821-719b-451b-9b19-4f1ae1b12994" containerName="init" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.235004 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b762821-719b-451b-9b19-4f1ae1b12994" containerName="init" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.235268 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="884b0d29-b0dd-48bf-bb02-51cbb4503b41" containerName="nova-cell1-conductor-db-sync" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.235300 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b263b3a-d693-45c2-8d18-ff0170ee5851" containerName="nova-manage" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.235335 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b762821-719b-451b-9b19-4f1ae1b12994" containerName="dnsmasq-dns" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.236367 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.242845 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.285108 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.360843 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjgnk\" (UniqueName: \"kubernetes.io/projected/cf11f1d4-587b-49a1-bd96-153799d2a6d4-kube-api-access-hjgnk\") pod \"nova-cell1-conductor-0\" (UID: \"cf11f1d4-587b-49a1-bd96-153799d2a6d4\") " pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.362926 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf11f1d4-587b-49a1-bd96-153799d2a6d4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cf11f1d4-587b-49a1-bd96-153799d2a6d4\") " pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.363910 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf11f1d4-587b-49a1-bd96-153799d2a6d4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cf11f1d4-587b-49a1-bd96-153799d2a6d4\") " pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.411662 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.412407 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerName="nova-api-api" containerID="cri-o://cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0" gracePeriod=30 Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.412720 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerName="nova-api-log" containerID="cri-o://b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8" gracePeriod=30 Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.448453 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.449520 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a92fd537-d2cb-4d18-843e-82a6671896ed" containerName="nova-scheduler-scheduler" containerID="cri-o://64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43" gracePeriod=30 Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.471214 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjgnk\" (UniqueName: \"kubernetes.io/projected/cf11f1d4-587b-49a1-bd96-153799d2a6d4-kube-api-access-hjgnk\") pod \"nova-cell1-conductor-0\" (UID: \"cf11f1d4-587b-49a1-bd96-153799d2a6d4\") " pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.471960 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf11f1d4-587b-49a1-bd96-153799d2a6d4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cf11f1d4-587b-49a1-bd96-153799d2a6d4\") " pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.472136 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf11f1d4-587b-49a1-bd96-153799d2a6d4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cf11f1d4-587b-49a1-bd96-153799d2a6d4\") " pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.505778 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjgnk\" (UniqueName: \"kubernetes.io/projected/cf11f1d4-587b-49a1-bd96-153799d2a6d4-kube-api-access-hjgnk\") pod \"nova-cell1-conductor-0\" (UID: \"cf11f1d4-587b-49a1-bd96-153799d2a6d4\") " pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.522373 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf11f1d4-587b-49a1-bd96-153799d2a6d4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cf11f1d4-587b-49a1-bd96-153799d2a6d4\") " pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.534694 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.535078 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="575a8ef8-bd25-4e2c-babe-36d73a352084" containerName="nova-metadata-log" containerID="cri-o://02dc8eed1072ba3e00495be50a98b0de98d4b6b2ff4e86c340f431447086a557" gracePeriod=30 Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.535671 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="575a8ef8-bd25-4e2c-babe-36d73a352084" containerName="nova-metadata-metadata" containerID="cri-o://66a361ce9ec5ba09b9840f06c1f868b0602bfa23b58c143e01de3a9f28322b1c" gracePeriod=30 Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.540375 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf11f1d4-587b-49a1-bd96-153799d2a6d4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cf11f1d4-587b-49a1-bd96-153799d2a6d4\") " pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:46 crc kubenswrapper[4954]: I1209 17:29:46.571509 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.111990 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 17:29:47 crc kubenswrapper[4954]: W1209 17:29:47.115889 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf11f1d4_587b_49a1_bd96_153799d2a6d4.slice/crio-3cfbe542ac778bf785404d63d561a266322f04e4343f4a4b0bfc7cd0102dff1a WatchSource:0}: Error finding container 3cfbe542ac778bf785404d63d561a266322f04e4343f4a4b0bfc7cd0102dff1a: Status 404 returned error can't find the container with id 3cfbe542ac778bf785404d63d561a266322f04e4343f4a4b0bfc7cd0102dff1a Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.205467 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cf11f1d4-587b-49a1-bd96-153799d2a6d4","Type":"ContainerStarted","Data":"3cfbe542ac778bf785404d63d561a266322f04e4343f4a4b0bfc7cd0102dff1a"} Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.211327 4954 generic.go:334] "Generic (PLEG): container finished" podID="575a8ef8-bd25-4e2c-babe-36d73a352084" containerID="66a361ce9ec5ba09b9840f06c1f868b0602bfa23b58c143e01de3a9f28322b1c" exitCode=0 Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.211363 4954 generic.go:334] "Generic (PLEG): container finished" podID="575a8ef8-bd25-4e2c-babe-36d73a352084" containerID="02dc8eed1072ba3e00495be50a98b0de98d4b6b2ff4e86c340f431447086a557" exitCode=143 Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.211417 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"575a8ef8-bd25-4e2c-babe-36d73a352084","Type":"ContainerDied","Data":"66a361ce9ec5ba09b9840f06c1f868b0602bfa23b58c143e01de3a9f28322b1c"} Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.211451 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"575a8ef8-bd25-4e2c-babe-36d73a352084","Type":"ContainerDied","Data":"02dc8eed1072ba3e00495be50a98b0de98d4b6b2ff4e86c340f431447086a557"} Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.247042 4954 generic.go:334] "Generic (PLEG): container finished" podID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerID="b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8" exitCode=143 Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.248422 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036dacf0-3f42-40cd-82d7-1b037268d95b","Type":"ContainerDied","Data":"b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8"} Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.403085 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.494853 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm7gd\" (UniqueName: \"kubernetes.io/projected/575a8ef8-bd25-4e2c-babe-36d73a352084-kube-api-access-zm7gd\") pod \"575a8ef8-bd25-4e2c-babe-36d73a352084\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.495128 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-nova-metadata-tls-certs\") pod \"575a8ef8-bd25-4e2c-babe-36d73a352084\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.495211 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-combined-ca-bundle\") pod \"575a8ef8-bd25-4e2c-babe-36d73a352084\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.495255 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/575a8ef8-bd25-4e2c-babe-36d73a352084-logs\") pod \"575a8ef8-bd25-4e2c-babe-36d73a352084\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.495300 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-config-data\") pod \"575a8ef8-bd25-4e2c-babe-36d73a352084\" (UID: \"575a8ef8-bd25-4e2c-babe-36d73a352084\") " Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.502056 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/575a8ef8-bd25-4e2c-babe-36d73a352084-logs" (OuterVolumeSpecName: "logs") pod "575a8ef8-bd25-4e2c-babe-36d73a352084" (UID: "575a8ef8-bd25-4e2c-babe-36d73a352084"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.545890 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/575a8ef8-bd25-4e2c-babe-36d73a352084-kube-api-access-zm7gd" (OuterVolumeSpecName: "kube-api-access-zm7gd") pod "575a8ef8-bd25-4e2c-babe-36d73a352084" (UID: "575a8ef8-bd25-4e2c-babe-36d73a352084"). InnerVolumeSpecName "kube-api-access-zm7gd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.572552 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-config-data" (OuterVolumeSpecName: "config-data") pod "575a8ef8-bd25-4e2c-babe-36d73a352084" (UID: "575a8ef8-bd25-4e2c-babe-36d73a352084"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.599090 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "575a8ef8-bd25-4e2c-babe-36d73a352084" (UID: "575a8ef8-bd25-4e2c-babe-36d73a352084"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.621096 4954 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/575a8ef8-bd25-4e2c-babe-36d73a352084-logs\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.621133 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.621146 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm7gd\" (UniqueName: \"kubernetes.io/projected/575a8ef8-bd25-4e2c-babe-36d73a352084-kube-api-access-zm7gd\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.621155 4954 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.635569 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "575a8ef8-bd25-4e2c-babe-36d73a352084" (UID: "575a8ef8-bd25-4e2c-babe-36d73a352084"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:47 crc kubenswrapper[4954]: I1209 17:29:47.733495 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575a8ef8-bd25-4e2c-babe-36d73a352084-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.259201 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cf11f1d4-587b-49a1-bd96-153799d2a6d4","Type":"ContainerStarted","Data":"c3b4a68d9c43bc5c37cccdfece77780871073d37e92d79a6820fe7b3b98d7b8f"} Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.259537 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.261913 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"575a8ef8-bd25-4e2c-babe-36d73a352084","Type":"ContainerDied","Data":"ee46cd55df0ec6b7af235716aeeeaed62407221624496f16c1c8d605e16317cf"} Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.261955 4954 scope.go:117] "RemoveContainer" containerID="66a361ce9ec5ba09b9840f06c1f868b0602bfa23b58c143e01de3a9f28322b1c" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.262008 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.287493 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.287471369 podStartE2EDuration="2.287471369s" podCreationTimestamp="2025-12-09 17:29:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:29:48.281854573 +0000 UTC m=+1984.670028393" watchObservedRunningTime="2025-12-09 17:29:48.287471369 +0000 UTC m=+1984.675645189" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.294170 4954 scope.go:117] "RemoveContainer" containerID="02dc8eed1072ba3e00495be50a98b0de98d4b6b2ff4e86c340f431447086a557" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.311031 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.322730 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.352091 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:48 crc kubenswrapper[4954]: E1209 17:29:48.352687 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="575a8ef8-bd25-4e2c-babe-36d73a352084" containerName="nova-metadata-log" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.352708 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="575a8ef8-bd25-4e2c-babe-36d73a352084" containerName="nova-metadata-log" Dec 09 17:29:48 crc kubenswrapper[4954]: E1209 17:29:48.352739 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="575a8ef8-bd25-4e2c-babe-36d73a352084" containerName="nova-metadata-metadata" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.352747 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="575a8ef8-bd25-4e2c-babe-36d73a352084" containerName="nova-metadata-metadata" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.353002 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="575a8ef8-bd25-4e2c-babe-36d73a352084" containerName="nova-metadata-log" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.353035 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="575a8ef8-bd25-4e2c-babe-36d73a352084" containerName="nova-metadata-metadata" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.355229 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.365401 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.365777 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.393905 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.460038 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-config-data\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.460100 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.460694 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmfnj\" (UniqueName: \"kubernetes.io/projected/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-kube-api-access-rmfnj\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.460803 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-logs\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.460861 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: E1209 17:29:48.535708 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 17:29:48 crc kubenswrapper[4954]: E1209 17:29:48.537424 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 17:29:48 crc kubenswrapper[4954]: E1209 17:29:48.539296 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 17:29:48 crc kubenswrapper[4954]: E1209 17:29:48.539390 4954 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a92fd537-d2cb-4d18-843e-82a6671896ed" containerName="nova-scheduler-scheduler" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.563369 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-config-data\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.563437 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.563537 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmfnj\" (UniqueName: \"kubernetes.io/projected/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-kube-api-access-rmfnj\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.563558 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-logs\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.563618 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.565219 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-logs\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.570288 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.570922 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.573932 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-config-data\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.581418 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmfnj\" (UniqueName: \"kubernetes.io/projected/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-kube-api-access-rmfnj\") pod \"nova-metadata-0\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " pod="openstack/nova-metadata-0" Dec 09 17:29:48 crc kubenswrapper[4954]: I1209 17:29:48.691448 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:29:49 crc kubenswrapper[4954]: W1209 17:29:49.232059 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8c4e70e_9b93_42c6_b2e7_ce97ecb4901c.slice/crio-263edc65ea2f453fe2957d20215351179a3a24b84e6bdeb955dfd0adafb0de2a WatchSource:0}: Error finding container 263edc65ea2f453fe2957d20215351179a3a24b84e6bdeb955dfd0adafb0de2a: Status 404 returned error can't find the container with id 263edc65ea2f453fe2957d20215351179a3a24b84e6bdeb955dfd0adafb0de2a Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.236829 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.275297 4954 generic.go:334] "Generic (PLEG): container finished" podID="f107a10b-0481-41b9-bcfa-7e5bef86ca1a" containerID="9bd5edc6798b8894974430538ce7f67ab22bb12c2d70de46d5e42f01fe505ae2" exitCode=0 Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.275365 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6q2z2" event={"ID":"f107a10b-0481-41b9-bcfa-7e5bef86ca1a","Type":"ContainerDied","Data":"9bd5edc6798b8894974430538ce7f67ab22bb12c2d70de46d5e42f01fe505ae2"} Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.286384 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c","Type":"ContainerStarted","Data":"263edc65ea2f453fe2957d20215351179a3a24b84e6bdeb955dfd0adafb0de2a"} Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.845725 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.901788 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-combined-ca-bundle\") pod \"a92fd537-d2cb-4d18-843e-82a6671896ed\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.901901 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-config-data\") pod \"a92fd537-d2cb-4d18-843e-82a6671896ed\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.901962 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpgjn\" (UniqueName: \"kubernetes.io/projected/a92fd537-d2cb-4d18-843e-82a6671896ed-kube-api-access-bpgjn\") pod \"a92fd537-d2cb-4d18-843e-82a6671896ed\" (UID: \"a92fd537-d2cb-4d18-843e-82a6671896ed\") " Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.912711 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a92fd537-d2cb-4d18-843e-82a6671896ed-kube-api-access-bpgjn" (OuterVolumeSpecName: "kube-api-access-bpgjn") pod "a92fd537-d2cb-4d18-843e-82a6671896ed" (UID: "a92fd537-d2cb-4d18-843e-82a6671896ed"). InnerVolumeSpecName "kube-api-access-bpgjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.965877 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-config-data" (OuterVolumeSpecName: "config-data") pod "a92fd537-d2cb-4d18-843e-82a6671896ed" (UID: "a92fd537-d2cb-4d18-843e-82a6671896ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:49 crc kubenswrapper[4954]: I1209 17:29:49.981758 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a92fd537-d2cb-4d18-843e-82a6671896ed" (UID: "a92fd537-d2cb-4d18-843e-82a6671896ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.005500 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.005548 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92fd537-d2cb-4d18-843e-82a6671896ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.005562 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpgjn\" (UniqueName: \"kubernetes.io/projected/a92fd537-d2cb-4d18-843e-82a6671896ed-kube-api-access-bpgjn\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.138405 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="575a8ef8-bd25-4e2c-babe-36d73a352084" path="/var/lib/kubelet/pods/575a8ef8-bd25-4e2c-babe-36d73a352084/volumes" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.305639 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.305860 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c","Type":"ContainerStarted","Data":"eb6a4c99ea7a347b0d2674943908b4ff21ab0ee5c6078de421b64f30dfcc7309"} Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.306161 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c","Type":"ContainerStarted","Data":"08f34ee49683df1db9d34a0f255a676999ad9be2ba72e974eec23eb81c93a0ba"} Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.308139 4954 generic.go:334] "Generic (PLEG): container finished" podID="a92fd537-d2cb-4d18-843e-82a6671896ed" containerID="64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43" exitCode=0 Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.308196 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.308202 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a92fd537-d2cb-4d18-843e-82a6671896ed","Type":"ContainerDied","Data":"64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43"} Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.308396 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a92fd537-d2cb-4d18-843e-82a6671896ed","Type":"ContainerDied","Data":"8fa72e19ca13c3843d3a670b601bf71ff8652e350e7a64d0696be0415282efd0"} Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.308457 4954 scope.go:117] "RemoveContainer" containerID="64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.318109 4954 generic.go:334] "Generic (PLEG): container finished" podID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerID="cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0" exitCode=0 Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.318254 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.318371 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036dacf0-3f42-40cd-82d7-1b037268d95b","Type":"ContainerDied","Data":"cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0"} Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.318407 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036dacf0-3f42-40cd-82d7-1b037268d95b","Type":"ContainerDied","Data":"8a09c2adcee8e4a39979fb264d2e53bc10f8dac38b0d33c60ebad5f113bd040a"} Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.349227 4954 scope.go:117] "RemoveContainer" containerID="64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43" Dec 09 17:29:50 crc kubenswrapper[4954]: E1209 17:29:50.353022 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43\": container with ID starting with 64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43 not found: ID does not exist" containerID="64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.353059 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43"} err="failed to get container status \"64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43\": rpc error: code = NotFound desc = could not find container \"64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43\": container with ID starting with 64740f9eb5ba121773f768152430671acf96f6d6f0d56f109a91ee9b38098c43 not found: ID does not exist" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.353084 4954 scope.go:117] "RemoveContainer" containerID="cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.388564 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.388536024 podStartE2EDuration="2.388536024s" podCreationTimestamp="2025-12-09 17:29:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:29:50.362116166 +0000 UTC m=+1986.750289986" watchObservedRunningTime="2025-12-09 17:29:50.388536024 +0000 UTC m=+1986.776709844" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.399369 4954 scope.go:117] "RemoveContainer" containerID="b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.402950 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.432796 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t42j6\" (UniqueName: \"kubernetes.io/projected/036dacf0-3f42-40cd-82d7-1b037268d95b-kube-api-access-t42j6\") pod \"036dacf0-3f42-40cd-82d7-1b037268d95b\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.432950 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-combined-ca-bundle\") pod \"036dacf0-3f42-40cd-82d7-1b037268d95b\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.433061 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036dacf0-3f42-40cd-82d7-1b037268d95b-logs\") pod \"036dacf0-3f42-40cd-82d7-1b037268d95b\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.433236 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-config-data\") pod \"036dacf0-3f42-40cd-82d7-1b037268d95b\" (UID: \"036dacf0-3f42-40cd-82d7-1b037268d95b\") " Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.449691 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.454439 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/036dacf0-3f42-40cd-82d7-1b037268d95b-kube-api-access-t42j6" (OuterVolumeSpecName: "kube-api-access-t42j6") pod "036dacf0-3f42-40cd-82d7-1b037268d95b" (UID: "036dacf0-3f42-40cd-82d7-1b037268d95b"). InnerVolumeSpecName "kube-api-access-t42j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.455235 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/036dacf0-3f42-40cd-82d7-1b037268d95b-logs" (OuterVolumeSpecName: "logs") pod "036dacf0-3f42-40cd-82d7-1b037268d95b" (UID: "036dacf0-3f42-40cd-82d7-1b037268d95b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.474150 4954 scope.go:117] "RemoveContainer" containerID="cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0" Dec 09 17:29:50 crc kubenswrapper[4954]: E1209 17:29:50.474812 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0\": container with ID starting with cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0 not found: ID does not exist" containerID="cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.474882 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0"} err="failed to get container status \"cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0\": rpc error: code = NotFound desc = could not find container \"cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0\": container with ID starting with cad68c50fe0bf86ffc118650e273cb625d887d46c7d226274ecd421ddf6b03e0 not found: ID does not exist" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.474916 4954 scope.go:117] "RemoveContainer" containerID="b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8" Dec 09 17:29:50 crc kubenswrapper[4954]: E1209 17:29:50.475326 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8\": container with ID starting with b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8 not found: ID does not exist" containerID="b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.476281 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8"} err="failed to get container status \"b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8\": rpc error: code = NotFound desc = could not find container \"b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8\": container with ID starting with b7524e6bb922a805c36caf3a17bdd7a50dcdbd6633fc55d02752ba8f612c0ee8 not found: ID does not exist" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.493793 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-config-data" (OuterVolumeSpecName: "config-data") pod "036dacf0-3f42-40cd-82d7-1b037268d95b" (UID: "036dacf0-3f42-40cd-82d7-1b037268d95b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.506721 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "036dacf0-3f42-40cd-82d7-1b037268d95b" (UID: "036dacf0-3f42-40cd-82d7-1b037268d95b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.508032 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:29:50 crc kubenswrapper[4954]: E1209 17:29:50.508572 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerName="nova-api-api" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.508595 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerName="nova-api-api" Dec 09 17:29:50 crc kubenswrapper[4954]: E1209 17:29:50.508669 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerName="nova-api-log" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.508676 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerName="nova-api-log" Dec 09 17:29:50 crc kubenswrapper[4954]: E1209 17:29:50.508701 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a92fd537-d2cb-4d18-843e-82a6671896ed" containerName="nova-scheduler-scheduler" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.508708 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a92fd537-d2cb-4d18-843e-82a6671896ed" containerName="nova-scheduler-scheduler" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.508948 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a92fd537-d2cb-4d18-843e-82a6671896ed" containerName="nova-scheduler-scheduler" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.508962 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerName="nova-api-log" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.508985 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" containerName="nova-api-api" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.510097 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.514007 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.529455 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.549094 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.549231 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-config-data\") pod \"nova-scheduler-0\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.549856 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lczf\" (UniqueName: \"kubernetes.io/projected/c2b4d601-8a86-48b0-abd3-039526cb6057-kube-api-access-7lczf\") pod \"nova-scheduler-0\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.550077 4954 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036dacf0-3f42-40cd-82d7-1b037268d95b-logs\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.550104 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.550121 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t42j6\" (UniqueName: \"kubernetes.io/projected/036dacf0-3f42-40cd-82d7-1b037268d95b-kube-api-access-t42j6\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.550136 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036dacf0-3f42-40cd-82d7-1b037268d95b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.652292 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lczf\" (UniqueName: \"kubernetes.io/projected/c2b4d601-8a86-48b0-abd3-039526cb6057-kube-api-access-7lczf\") pod \"nova-scheduler-0\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.652406 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.652553 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-config-data\") pod \"nova-scheduler-0\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.665957 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-config-data\") pod \"nova-scheduler-0\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.670871 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.673357 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lczf\" (UniqueName: \"kubernetes.io/projected/c2b4d601-8a86-48b0-abd3-039526cb6057-kube-api-access-7lczf\") pod \"nova-scheduler-0\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.691695 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.701506 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.710954 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.716207 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.729239 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.747399 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.756793 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-config-data\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.756868 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.756908 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6prqn\" (UniqueName: \"kubernetes.io/projected/af2878c3-2924-452b-93a9-0a2a843cfdff-kube-api-access-6prqn\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.757100 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af2878c3-2924-452b-93a9-0a2a843cfdff-logs\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.835027 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.850701 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.860029 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-config-data\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.860089 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.860126 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6prqn\" (UniqueName: \"kubernetes.io/projected/af2878c3-2924-452b-93a9-0a2a843cfdff-kube-api-access-6prqn\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.860309 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af2878c3-2924-452b-93a9-0a2a843cfdff-logs\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.860858 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af2878c3-2924-452b-93a9-0a2a843cfdff-logs\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.872354 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-config-data\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.873085 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:50 crc kubenswrapper[4954]: I1209 17:29:50.885172 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6prqn\" (UniqueName: \"kubernetes.io/projected/af2878c3-2924-452b-93a9-0a2a843cfdff-kube-api-access-6prqn\") pod \"nova-api-0\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " pod="openstack/nova-api-0" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:50.962022 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7zr8\" (UniqueName: \"kubernetes.io/projected/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-kube-api-access-s7zr8\") pod \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:50.962144 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-config-data\") pod \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:50.962303 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-scripts\") pod \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:50.962359 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-combined-ca-bundle\") pod \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\" (UID: \"f107a10b-0481-41b9-bcfa-7e5bef86ca1a\") " Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.015911 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-kube-api-access-s7zr8" (OuterVolumeSpecName: "kube-api-access-s7zr8") pod "f107a10b-0481-41b9-bcfa-7e5bef86ca1a" (UID: "f107a10b-0481-41b9-bcfa-7e5bef86ca1a"). InnerVolumeSpecName "kube-api-access-s7zr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.021156 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-scripts" (OuterVolumeSpecName: "scripts") pod "f107a10b-0481-41b9-bcfa-7e5bef86ca1a" (UID: "f107a10b-0481-41b9-bcfa-7e5bef86ca1a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.048982 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f107a10b-0481-41b9-bcfa-7e5bef86ca1a" (UID: "f107a10b-0481-41b9-bcfa-7e5bef86ca1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.065337 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.065384 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.065402 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7zr8\" (UniqueName: \"kubernetes.io/projected/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-kube-api-access-s7zr8\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.073331 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-config-data" (OuterVolumeSpecName: "config-data") pod "f107a10b-0481-41b9-bcfa-7e5bef86ca1a" (UID: "f107a10b-0481-41b9-bcfa-7e5bef86ca1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.144220 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.168709 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f107a10b-0481-41b9-bcfa-7e5bef86ca1a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.336646 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6q2z2" event={"ID":"f107a10b-0481-41b9-bcfa-7e5bef86ca1a","Type":"ContainerDied","Data":"af0feae622680ba4f2085c6b0360cb05a412b1d4ee8fa3549bf822472fe12aa2"} Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.337031 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af0feae622680ba4f2085c6b0360cb05a412b1d4ee8fa3549bf822472fe12aa2" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.336757 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6q2z2" Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.389453 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:29:51 crc kubenswrapper[4954]: I1209 17:29:51.634610 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.139222 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="036dacf0-3f42-40cd-82d7-1b037268d95b" path="/var/lib/kubelet/pods/036dacf0-3f42-40cd-82d7-1b037268d95b/volumes" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.141498 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a92fd537-d2cb-4d18-843e-82a6671896ed" path="/var/lib/kubelet/pods/a92fd537-d2cb-4d18-843e-82a6671896ed/volumes" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.351308 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"af2878c3-2924-452b-93a9-0a2a843cfdff","Type":"ContainerStarted","Data":"ead158534d1701a55c24f0c87725e83d27fddaa15c59187aebe22aa92e53e15e"} Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.351356 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"af2878c3-2924-452b-93a9-0a2a843cfdff","Type":"ContainerStarted","Data":"49603484aa52c31a7f74ce4ecf682f4d2a0fc175479035cb29c2e14aa10e9cf5"} Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.351367 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"af2878c3-2924-452b-93a9-0a2a843cfdff","Type":"ContainerStarted","Data":"474eaf2f7919a9c6cd02f00b06cc25b07099cc66f72edcb9453c487fc21e0eed"} Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.357040 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c2b4d601-8a86-48b0-abd3-039526cb6057","Type":"ContainerStarted","Data":"fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1"} Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.357102 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c2b4d601-8a86-48b0-abd3-039526cb6057","Type":"ContainerStarted","Data":"2aec6d21390d884799e01e04b960cdc11a2eb785e727527682acebe04e80647c"} Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.376423 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.376396445 podStartE2EDuration="2.376396445s" podCreationTimestamp="2025-12-09 17:29:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:29:52.371663066 +0000 UTC m=+1988.759836906" watchObservedRunningTime="2025-12-09 17:29:52.376396445 +0000 UTC m=+1988.764570265" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.405849 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.4058127750000002 podStartE2EDuration="2.405812775s" podCreationTimestamp="2025-12-09 17:29:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:29:52.394517532 +0000 UTC m=+1988.782691352" watchObservedRunningTime="2025-12-09 17:29:52.405812775 +0000 UTC m=+1988.793986615" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.545582 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 09 17:29:52 crc kubenswrapper[4954]: E1209 17:29:52.546450 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f107a10b-0481-41b9-bcfa-7e5bef86ca1a" containerName="aodh-db-sync" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.546527 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f107a10b-0481-41b9-bcfa-7e5bef86ca1a" containerName="aodh-db-sync" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.546839 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f107a10b-0481-41b9-bcfa-7e5bef86ca1a" containerName="aodh-db-sync" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.572627 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.595421 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.595887 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-hh57v" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.596005 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.663920 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.713489 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-config-data\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.713662 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xxj2\" (UniqueName: \"kubernetes.io/projected/93f122ae-71b4-4393-a163-9f61a9b52b8d-kube-api-access-4xxj2\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.713721 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-scripts\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.713739 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.815900 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xxj2\" (UniqueName: \"kubernetes.io/projected/93f122ae-71b4-4393-a163-9f61a9b52b8d-kube-api-access-4xxj2\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.816659 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-scripts\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.816694 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.816836 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-config-data\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.825877 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-scripts\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.826299 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-config-data\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.831259 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.847470 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xxj2\" (UniqueName: \"kubernetes.io/projected/93f122ae-71b4-4393-a163-9f61a9b52b8d-kube-api-access-4xxj2\") pod \"aodh-0\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " pod="openstack/aodh-0" Dec 09 17:29:52 crc kubenswrapper[4954]: I1209 17:29:52.987674 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 17:29:53 crc kubenswrapper[4954]: I1209 17:29:53.592767 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 09 17:29:53 crc kubenswrapper[4954]: W1209 17:29:53.608230 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93f122ae_71b4_4393_a163_9f61a9b52b8d.slice/crio-0db99a42c5a92ef8cc6d078530dff14e39a64e18f80f92f595498a8de558c5f8 WatchSource:0}: Error finding container 0db99a42c5a92ef8cc6d078530dff14e39a64e18f80f92f595498a8de558c5f8: Status 404 returned error can't find the container with id 0db99a42c5a92ef8cc6d078530dff14e39a64e18f80f92f595498a8de558c5f8 Dec 09 17:29:53 crc kubenswrapper[4954]: I1209 17:29:53.692536 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 17:29:53 crc kubenswrapper[4954]: I1209 17:29:53.692618 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 17:29:54 crc kubenswrapper[4954]: I1209 17:29:54.381259 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"93f122ae-71b4-4393-a163-9f61a9b52b8d","Type":"ContainerStarted","Data":"0db99a42c5a92ef8cc6d078530dff14e39a64e18f80f92f595498a8de558c5f8"} Dec 09 17:29:54 crc kubenswrapper[4954]: I1209 17:29:54.996311 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:54 crc kubenswrapper[4954]: I1209 17:29:54.996940 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="ceilometer-central-agent" containerID="cri-o://061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80" gracePeriod=30 Dec 09 17:29:54 crc kubenswrapper[4954]: I1209 17:29:54.996998 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="sg-core" containerID="cri-o://bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824" gracePeriod=30 Dec 09 17:29:54 crc kubenswrapper[4954]: I1209 17:29:54.997042 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="ceilometer-notification-agent" containerID="cri-o://1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb" gracePeriod=30 Dec 09 17:29:54 crc kubenswrapper[4954]: I1209 17:29:54.997370 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="proxy-httpd" containerID="cri-o://969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8" gracePeriod=30 Dec 09 17:29:55 crc kubenswrapper[4954]: I1209 17:29:55.010363 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.229:3000/\": EOF" Dec 09 17:29:55 crc kubenswrapper[4954]: I1209 17:29:55.397039 4954 generic.go:334] "Generic (PLEG): container finished" podID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerID="969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8" exitCode=0 Dec 09 17:29:55 crc kubenswrapper[4954]: I1209 17:29:55.397075 4954 generic.go:334] "Generic (PLEG): container finished" podID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerID="bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824" exitCode=2 Dec 09 17:29:55 crc kubenswrapper[4954]: I1209 17:29:55.397123 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f","Type":"ContainerDied","Data":"969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8"} Dec 09 17:29:55 crc kubenswrapper[4954]: I1209 17:29:55.397158 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f","Type":"ContainerDied","Data":"bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824"} Dec 09 17:29:55 crc kubenswrapper[4954]: I1209 17:29:55.398959 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"93f122ae-71b4-4393-a163-9f61a9b52b8d","Type":"ContainerStarted","Data":"d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8"} Dec 09 17:29:55 crc kubenswrapper[4954]: I1209 17:29:55.836159 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 17:29:55 crc kubenswrapper[4954]: I1209 17:29:55.884054 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 09 17:29:56 crc kubenswrapper[4954]: I1209 17:29:56.436882 4954 generic.go:334] "Generic (PLEG): container finished" podID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerID="061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80" exitCode=0 Dec 09 17:29:56 crc kubenswrapper[4954]: I1209 17:29:56.437087 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f","Type":"ContainerDied","Data":"061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80"} Dec 09 17:29:56 crc kubenswrapper[4954]: I1209 17:29:56.613823 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.327015 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.436184 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-sg-core-conf-yaml\") pod \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.436386 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-scripts\") pod \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.436429 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-log-httpd\") pod \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.436570 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tw7z\" (UniqueName: \"kubernetes.io/projected/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-kube-api-access-7tw7z\") pod \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.436652 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-config-data\") pod \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.436701 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-combined-ca-bundle\") pod \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.436737 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-run-httpd\") pod \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\" (UID: \"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f\") " Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.437529 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" (UID: "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.437785 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" (UID: "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.442726 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-kube-api-access-7tw7z" (OuterVolumeSpecName: "kube-api-access-7tw7z") pod "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" (UID: "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f"). InnerVolumeSpecName "kube-api-access-7tw7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.443021 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-scripts" (OuterVolumeSpecName: "scripts") pod "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" (UID: "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.461188 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"93f122ae-71b4-4393-a163-9f61a9b52b8d","Type":"ContainerStarted","Data":"fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f"} Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.471056 4954 generic.go:334] "Generic (PLEG): container finished" podID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerID="1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb" exitCode=0 Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.471131 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f","Type":"ContainerDied","Data":"1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb"} Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.471160 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.471210 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f","Type":"ContainerDied","Data":"c31b450ecd1b3089c6b9e26dfa1b411256a71d50bbc51c43f085199f75def060"} Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.471235 4954 scope.go:117] "RemoveContainer" containerID="969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.489468 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" (UID: "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.509609 4954 scope.go:117] "RemoveContainer" containerID="bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.539759 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tw7z\" (UniqueName: \"kubernetes.io/projected/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-kube-api-access-7tw7z\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.539794 4954 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.539807 4954 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.539818 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.539829 4954 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.542841 4954 scope.go:117] "RemoveContainer" containerID="1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.555724 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" (UID: "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.575297 4954 scope.go:117] "RemoveContainer" containerID="061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.596173 4954 scope.go:117] "RemoveContainer" containerID="969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8" Dec 09 17:29:57 crc kubenswrapper[4954]: E1209 17:29:57.597108 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8\": container with ID starting with 969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8 not found: ID does not exist" containerID="969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.597187 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8"} err="failed to get container status \"969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8\": rpc error: code = NotFound desc = could not find container \"969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8\": container with ID starting with 969b67fb3964c878393c3321ad3c55af92384562ef7b5ef7850cf4261e3894c8 not found: ID does not exist" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.597213 4954 scope.go:117] "RemoveContainer" containerID="bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824" Dec 09 17:29:57 crc kubenswrapper[4954]: E1209 17:29:57.597904 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824\": container with ID starting with bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824 not found: ID does not exist" containerID="bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.597944 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824"} err="failed to get container status \"bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824\": rpc error: code = NotFound desc = could not find container \"bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824\": container with ID starting with bc81d889b1f861ef2f04c17d08fb0a49d2cd2002765dfa86d070dc8ae6f24824 not found: ID does not exist" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.597961 4954 scope.go:117] "RemoveContainer" containerID="1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb" Dec 09 17:29:57 crc kubenswrapper[4954]: E1209 17:29:57.598272 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb\": container with ID starting with 1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb not found: ID does not exist" containerID="1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.598292 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb"} err="failed to get container status \"1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb\": rpc error: code = NotFound desc = could not find container \"1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb\": container with ID starting with 1565a97103ece6cfebc9b7d7f0c34ea92fbb489ef34de09da153da655f3d3fdb not found: ID does not exist" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.598305 4954 scope.go:117] "RemoveContainer" containerID="061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80" Dec 09 17:29:57 crc kubenswrapper[4954]: E1209 17:29:57.598718 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80\": container with ID starting with 061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80 not found: ID does not exist" containerID="061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.598759 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80"} err="failed to get container status \"061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80\": rpc error: code = NotFound desc = could not find container \"061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80\": container with ID starting with 061438d8a4f26cbefc93ca8a492b19147017ae0d4e2d1e1034162e76afa99d80 not found: ID does not exist" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.605887 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-config-data" (OuterVolumeSpecName: "config-data") pod "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" (UID: "cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.641696 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.641732 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.870131 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.885063 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.907336 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:57 crc kubenswrapper[4954]: E1209 17:29:57.908105 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="proxy-httpd" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.908134 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="proxy-httpd" Dec 09 17:29:57 crc kubenswrapper[4954]: E1209 17:29:57.908168 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="sg-core" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.908178 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="sg-core" Dec 09 17:29:57 crc kubenswrapper[4954]: E1209 17:29:57.908200 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="ceilometer-central-agent" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.908211 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="ceilometer-central-agent" Dec 09 17:29:57 crc kubenswrapper[4954]: E1209 17:29:57.908245 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="ceilometer-notification-agent" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.908254 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="ceilometer-notification-agent" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.908562 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="proxy-httpd" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.908621 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="ceilometer-notification-agent" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.908641 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="sg-core" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.908659 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" containerName="ceilometer-central-agent" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.913117 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.942859 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.942873 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 17:29:57 crc kubenswrapper[4954]: I1209 17:29:57.946828 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.051685 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.052314 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-scripts\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.052463 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.052547 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-config-data\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.052775 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-run-httpd\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.053161 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-log-httpd\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.053405 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjvkt\" (UniqueName: \"kubernetes.io/projected/11c3d434-a778-48db-87d6-d96942b8578a-kube-api-access-fjvkt\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.142081 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f" path="/var/lib/kubelet/pods/cd8c7f09-4d10-44d7-ad72-9ba2842c9b5f/volumes" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.158654 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-scripts\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.159035 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.159143 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-config-data\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.159281 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-run-httpd\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.159492 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-log-httpd\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.159641 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjvkt\" (UniqueName: \"kubernetes.io/projected/11c3d434-a778-48db-87d6-d96942b8578a-kube-api-access-fjvkt\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.159796 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.159837 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-run-httpd\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.160112 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-log-httpd\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.166052 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.166258 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.181294 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-scripts\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.181380 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjvkt\" (UniqueName: \"kubernetes.io/projected/11c3d434-a778-48db-87d6-d96942b8578a-kube-api-access-fjvkt\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.182043 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-config-data\") pod \"ceilometer-0\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.266016 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.692870 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.693218 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 17:29:58 crc kubenswrapper[4954]: I1209 17:29:58.791013 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:29:58 crc kubenswrapper[4954]: W1209 17:29:58.850062 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11c3d434_a778_48db_87d6_d96942b8578a.slice/crio-21e5095dacffc25c5ea7bb8bab538adb79138deb3a7692509709937bcbcbbe2c WatchSource:0}: Error finding container 21e5095dacffc25c5ea7bb8bab538adb79138deb3a7692509709937bcbcbbe2c: Status 404 returned error can't find the container with id 21e5095dacffc25c5ea7bb8bab538adb79138deb3a7692509709937bcbcbbe2c Dec 09 17:29:59 crc kubenswrapper[4954]: I1209 17:29:59.507048 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"93f122ae-71b4-4393-a163-9f61a9b52b8d","Type":"ContainerStarted","Data":"3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e"} Dec 09 17:29:59 crc kubenswrapper[4954]: I1209 17:29:59.508262 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c3d434-a778-48db-87d6-d96942b8578a","Type":"ContainerStarted","Data":"21e5095dacffc25c5ea7bb8bab538adb79138deb3a7692509709937bcbcbbe2c"} Dec 09 17:29:59 crc kubenswrapper[4954]: I1209 17:29:59.713879 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.233:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 17:29:59 crc kubenswrapper[4954]: I1209 17:29:59.713883 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.233:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.161625 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s"] Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.164028 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.168237 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.168491 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.198066 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s"] Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.318631 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f789cd82-95d6-4192-b4cf-586ff93babcf-secret-volume\") pod \"collect-profiles-29421690-dkg8s\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.318685 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhllc\" (UniqueName: \"kubernetes.io/projected/f789cd82-95d6-4192-b4cf-586ff93babcf-kube-api-access-vhllc\") pod \"collect-profiles-29421690-dkg8s\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.318721 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f789cd82-95d6-4192-b4cf-586ff93babcf-config-volume\") pod \"collect-profiles-29421690-dkg8s\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.421026 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f789cd82-95d6-4192-b4cf-586ff93babcf-config-volume\") pod \"collect-profiles-29421690-dkg8s\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.421808 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f789cd82-95d6-4192-b4cf-586ff93babcf-secret-volume\") pod \"collect-profiles-29421690-dkg8s\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.421918 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhllc\" (UniqueName: \"kubernetes.io/projected/f789cd82-95d6-4192-b4cf-586ff93babcf-kube-api-access-vhllc\") pod \"collect-profiles-29421690-dkg8s\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.423308 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f789cd82-95d6-4192-b4cf-586ff93babcf-config-volume\") pod \"collect-profiles-29421690-dkg8s\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.427246 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f789cd82-95d6-4192-b4cf-586ff93babcf-secret-volume\") pod \"collect-profiles-29421690-dkg8s\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.442184 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhllc\" (UniqueName: \"kubernetes.io/projected/f789cd82-95d6-4192-b4cf-586ff93babcf-kube-api-access-vhllc\") pod \"collect-profiles-29421690-dkg8s\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.512511 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.537374 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c3d434-a778-48db-87d6-d96942b8578a","Type":"ContainerStarted","Data":"4e02788a4e2e3301d0f021f658a3d4525948d82ae4c64555a7ba09ef764c5e97"} Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.836437 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 17:30:00 crc kubenswrapper[4954]: I1209 17:30:00.878930 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 17:30:01 crc kubenswrapper[4954]: I1209 17:30:01.145739 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 17:30:01 crc kubenswrapper[4954]: I1209 17:30:01.146104 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 17:30:01 crc kubenswrapper[4954]: I1209 17:30:01.271836 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s"] Dec 09 17:30:01 crc kubenswrapper[4954]: W1209 17:30:01.360204 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf789cd82_95d6_4192_b4cf_586ff93babcf.slice/crio-5860ff4b130b4e0812b305b53ef9ff4040522eac93685452190f3a85934bc5a7 WatchSource:0}: Error finding container 5860ff4b130b4e0812b305b53ef9ff4040522eac93685452190f3a85934bc5a7: Status 404 returned error can't find the container with id 5860ff4b130b4e0812b305b53ef9ff4040522eac93685452190f3a85934bc5a7 Dec 09 17:30:01 crc kubenswrapper[4954]: I1209 17:30:01.560774 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" event={"ID":"f789cd82-95d6-4192-b4cf-586ff93babcf","Type":"ContainerStarted","Data":"5860ff4b130b4e0812b305b53ef9ff4040522eac93685452190f3a85934bc5a7"} Dec 09 17:30:01 crc kubenswrapper[4954]: I1209 17:30:01.608876 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.227859 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.235:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.227911 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.235:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.580762 4954 generic.go:334] "Generic (PLEG): container finished" podID="f789cd82-95d6-4192-b4cf-586ff93babcf" containerID="0673e594c40666bba9c0a3d7507149040669bdd1282cf18fa4e895d6d81b87b5" exitCode=0 Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.581085 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" event={"ID":"f789cd82-95d6-4192-b4cf-586ff93babcf","Type":"ContainerDied","Data":"0673e594c40666bba9c0a3d7507149040669bdd1282cf18fa4e895d6d81b87b5"} Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.586391 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"93f122ae-71b4-4393-a163-9f61a9b52b8d","Type":"ContainerStarted","Data":"ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491"} Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.586560 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-api" containerID="cri-o://d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8" gracePeriod=30 Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.586754 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-listener" containerID="cri-o://ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491" gracePeriod=30 Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.586798 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-notifier" containerID="cri-o://3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e" gracePeriod=30 Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.586830 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-evaluator" containerID="cri-o://fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f" gracePeriod=30 Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.599574 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c3d434-a778-48db-87d6-d96942b8578a","Type":"ContainerStarted","Data":"3aaba52528c9b989ba9cfc2bcf2d55d164d167b06f23f7a58e435e19b0d98544"} Dec 09 17:30:02 crc kubenswrapper[4954]: I1209 17:30:02.651010 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.735795777 podStartE2EDuration="10.650979646s" podCreationTimestamp="2025-12-09 17:29:52 +0000 UTC" firstStartedPulling="2025-12-09 17:29:53.613387279 +0000 UTC m=+1990.001561099" lastFinishedPulling="2025-12-09 17:30:01.528571138 +0000 UTC m=+1997.916744968" observedRunningTime="2025-12-09 17:30:02.636865934 +0000 UTC m=+1999.025039764" watchObservedRunningTime="2025-12-09 17:30:02.650979646 +0000 UTC m=+1999.039153466" Dec 09 17:30:03 crc kubenswrapper[4954]: I1209 17:30:03.615932 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c3d434-a778-48db-87d6-d96942b8578a","Type":"ContainerStarted","Data":"3409e8d38448d1e74a39cc00ac58406a384efa8ebce20565943fb4118fa1f483"} Dec 09 17:30:03 crc kubenswrapper[4954]: I1209 17:30:03.618158 4954 generic.go:334] "Generic (PLEG): container finished" podID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerID="fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f" exitCode=0 Dec 09 17:30:03 crc kubenswrapper[4954]: I1209 17:30:03.618196 4954 generic.go:334] "Generic (PLEG): container finished" podID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerID="d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8" exitCode=0 Dec 09 17:30:03 crc kubenswrapper[4954]: I1209 17:30:03.618468 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"93f122ae-71b4-4393-a163-9f61a9b52b8d","Type":"ContainerDied","Data":"fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f"} Dec 09 17:30:03 crc kubenswrapper[4954]: I1209 17:30:03.618502 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"93f122ae-71b4-4393-a163-9f61a9b52b8d","Type":"ContainerDied","Data":"d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8"} Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.095520 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.232850 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f789cd82-95d6-4192-b4cf-586ff93babcf-secret-volume\") pod \"f789cd82-95d6-4192-b4cf-586ff93babcf\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.233173 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhllc\" (UniqueName: \"kubernetes.io/projected/f789cd82-95d6-4192-b4cf-586ff93babcf-kube-api-access-vhllc\") pod \"f789cd82-95d6-4192-b4cf-586ff93babcf\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.233250 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f789cd82-95d6-4192-b4cf-586ff93babcf-config-volume\") pod \"f789cd82-95d6-4192-b4cf-586ff93babcf\" (UID: \"f789cd82-95d6-4192-b4cf-586ff93babcf\") " Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.234699 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f789cd82-95d6-4192-b4cf-586ff93babcf-config-volume" (OuterVolumeSpecName: "config-volume") pod "f789cd82-95d6-4192-b4cf-586ff93babcf" (UID: "f789cd82-95d6-4192-b4cf-586ff93babcf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.258986 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f789cd82-95d6-4192-b4cf-586ff93babcf-kube-api-access-vhllc" (OuterVolumeSpecName: "kube-api-access-vhllc") pod "f789cd82-95d6-4192-b4cf-586ff93babcf" (UID: "f789cd82-95d6-4192-b4cf-586ff93babcf"). InnerVolumeSpecName "kube-api-access-vhllc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.268214 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f789cd82-95d6-4192-b4cf-586ff93babcf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f789cd82-95d6-4192-b4cf-586ff93babcf" (UID: "f789cd82-95d6-4192-b4cf-586ff93babcf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.336816 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhllc\" (UniqueName: \"kubernetes.io/projected/f789cd82-95d6-4192-b4cf-586ff93babcf-kube-api-access-vhllc\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.336860 4954 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f789cd82-95d6-4192-b4cf-586ff93babcf-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.336872 4954 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f789cd82-95d6-4192-b4cf-586ff93babcf-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.632735 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.632734 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s" event={"ID":"f789cd82-95d6-4192-b4cf-586ff93babcf","Type":"ContainerDied","Data":"5860ff4b130b4e0812b305b53ef9ff4040522eac93685452190f3a85934bc5a7"} Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.632792 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5860ff4b130b4e0812b305b53ef9ff4040522eac93685452190f3a85934bc5a7" Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.636316 4954 generic.go:334] "Generic (PLEG): container finished" podID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerID="3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e" exitCode=0 Dec 09 17:30:04 crc kubenswrapper[4954]: I1209 17:30:04.636364 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"93f122ae-71b4-4393-a163-9f61a9b52b8d","Type":"ContainerDied","Data":"3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e"} Dec 09 17:30:05 crc kubenswrapper[4954]: I1209 17:30:05.203711 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9"] Dec 09 17:30:05 crc kubenswrapper[4954]: I1209 17:30:05.208887 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421645-tk9h9"] Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.140997 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81" path="/var/lib/kubelet/pods/c6fdcef1-fc3f-4be1-a0f1-d68c73cc4c81/volumes" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.233096 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.388582 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-combined-ca-bundle\") pod \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.388656 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-config-data\") pod \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.388896 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnddt\" (UniqueName: \"kubernetes.io/projected/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-kube-api-access-nnddt\") pod \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\" (UID: \"5b4bfb0c-c99c-4c69-bc82-586084c33d2a\") " Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.394730 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-kube-api-access-nnddt" (OuterVolumeSpecName: "kube-api-access-nnddt") pod "5b4bfb0c-c99c-4c69-bc82-586084c33d2a" (UID: "5b4bfb0c-c99c-4c69-bc82-586084c33d2a"). InnerVolumeSpecName "kube-api-access-nnddt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.424357 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-config-data" (OuterVolumeSpecName: "config-data") pod "5b4bfb0c-c99c-4c69-bc82-586084c33d2a" (UID: "5b4bfb0c-c99c-4c69-bc82-586084c33d2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.430341 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b4bfb0c-c99c-4c69-bc82-586084c33d2a" (UID: "5b4bfb0c-c99c-4c69-bc82-586084c33d2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.492432 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.492472 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.492484 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnddt\" (UniqueName: \"kubernetes.io/projected/5b4bfb0c-c99c-4c69-bc82-586084c33d2a-kube-api-access-nnddt\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.666137 4954 generic.go:334] "Generic (PLEG): container finished" podID="5b4bfb0c-c99c-4c69-bc82-586084c33d2a" containerID="f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090" exitCode=137 Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.666200 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5b4bfb0c-c99c-4c69-bc82-586084c33d2a","Type":"ContainerDied","Data":"f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090"} Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.666209 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.666228 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5b4bfb0c-c99c-4c69-bc82-586084c33d2a","Type":"ContainerDied","Data":"5e900716bf1eb273d5f31dc027ab6f2d3bd45843398caecaa6b7308f2b7b22d8"} Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.666250 4954 scope.go:117] "RemoveContainer" containerID="f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.712815 4954 scope.go:117] "RemoveContainer" containerID="f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090" Dec 09 17:30:06 crc kubenswrapper[4954]: E1209 17:30:06.713499 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090\": container with ID starting with f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090 not found: ID does not exist" containerID="f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.713587 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090"} err="failed to get container status \"f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090\": rpc error: code = NotFound desc = could not find container \"f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090\": container with ID starting with f194590f15fe1e4546303042a19df0e2df758b484a0f1949253067cf86d07090 not found: ID does not exist" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.713669 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.724832 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.747871 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 17:30:06 crc kubenswrapper[4954]: E1209 17:30:06.748668 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4bfb0c-c99c-4c69-bc82-586084c33d2a" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.748698 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4bfb0c-c99c-4c69-bc82-586084c33d2a" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 17:30:06 crc kubenswrapper[4954]: E1209 17:30:06.748721 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f789cd82-95d6-4192-b4cf-586ff93babcf" containerName="collect-profiles" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.748730 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f789cd82-95d6-4192-b4cf-586ff93babcf" containerName="collect-profiles" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.749019 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4bfb0c-c99c-4c69-bc82-586084c33d2a" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.749065 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f789cd82-95d6-4192-b4cf-586ff93babcf" containerName="collect-profiles" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.749993 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.753081 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.753390 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.755612 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.760674 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.902451 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.903085 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.903190 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.903229 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:06 crc kubenswrapper[4954]: I1209 17:30:06.903266 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znm4z\" (UniqueName: \"kubernetes.io/projected/29eaf432-7585-424b-9732-f7e0bb49f380-kube-api-access-znm4z\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.005357 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.005472 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.005499 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.005525 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znm4z\" (UniqueName: \"kubernetes.io/projected/29eaf432-7585-424b-9732-f7e0bb49f380-kube-api-access-znm4z\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.005640 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.009817 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.010464 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.010893 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.020981 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/29eaf432-7585-424b-9732-f7e0bb49f380-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.023838 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znm4z\" (UniqueName: \"kubernetes.io/projected/29eaf432-7585-424b-9732-f7e0bb49f380-kube-api-access-znm4z\") pod \"nova-cell1-novncproxy-0\" (UID: \"29eaf432-7585-424b-9732-f7e0bb49f380\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.100092 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.593518 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 17:30:07 crc kubenswrapper[4954]: W1209 17:30:07.606843 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29eaf432_7585_424b_9732_f7e0bb49f380.slice/crio-5e7e23d84a0f573d584be861f04e1a80c65b1596e68ae03655a78533e23b26fe WatchSource:0}: Error finding container 5e7e23d84a0f573d584be861f04e1a80c65b1596e68ae03655a78533e23b26fe: Status 404 returned error can't find the container with id 5e7e23d84a0f573d584be861f04e1a80c65b1596e68ae03655a78533e23b26fe Dec 09 17:30:07 crc kubenswrapper[4954]: I1209 17:30:07.680151 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"29eaf432-7585-424b-9732-f7e0bb49f380","Type":"ContainerStarted","Data":"5e7e23d84a0f573d584be861f04e1a80c65b1596e68ae03655a78533e23b26fe"} Dec 09 17:30:08 crc kubenswrapper[4954]: I1209 17:30:08.134428 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b4bfb0c-c99c-4c69-bc82-586084c33d2a" path="/var/lib/kubelet/pods/5b4bfb0c-c99c-4c69-bc82-586084c33d2a/volumes" Dec 09 17:30:08 crc kubenswrapper[4954]: I1209 17:30:08.696171 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"29eaf432-7585-424b-9732-f7e0bb49f380","Type":"ContainerStarted","Data":"c12ddbfcdd1ec327ef13942f8ff1193e1a5b1db84b4c3baf3a44fdcc74ee2d99"} Dec 09 17:30:08 crc kubenswrapper[4954]: I1209 17:30:08.700318 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 17:30:08 crc kubenswrapper[4954]: I1209 17:30:08.702683 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 17:30:08 crc kubenswrapper[4954]: I1209 17:30:08.713257 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 17:30:08 crc kubenswrapper[4954]: I1209 17:30:08.731072 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.7310489650000003 podStartE2EDuration="2.731048965s" podCreationTimestamp="2025-12-09 17:30:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:30:08.719097501 +0000 UTC m=+2005.107271321" watchObservedRunningTime="2025-12-09 17:30:08.731048965 +0000 UTC m=+2005.119222775" Dec 09 17:30:09 crc kubenswrapper[4954]: I1209 17:30:09.717620 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.149079 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.149157 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.149922 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.149945 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.153526 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.153916 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.392657 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl"] Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.395038 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.432852 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl"] Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.515277 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.516739 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.516895 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-config\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.517016 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c86ql\" (UniqueName: \"kubernetes.io/projected/507b400c-fcae-4257-a6e5-2485da532750-kube-api-access-c86ql\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.517302 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.517443 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.619861 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.619925 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.619960 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-config\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.619991 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c86ql\" (UniqueName: \"kubernetes.io/projected/507b400c-fcae-4257-a6e5-2485da532750-kube-api-access-c86ql\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.620058 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.620100 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.621150 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.621162 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-config\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.621191 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.621158 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.621237 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.652621 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c86ql\" (UniqueName: \"kubernetes.io/projected/507b400c-fcae-4257-a6e5-2485da532750-kube-api-access-c86ql\") pod \"dnsmasq-dns-6b7bbf7cf9-kw4bl\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:11 crc kubenswrapper[4954]: I1209 17:30:11.726567 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:12 crc kubenswrapper[4954]: I1209 17:30:12.100568 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:12 crc kubenswrapper[4954]: I1209 17:30:12.234518 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl"] Dec 09 17:30:12 crc kubenswrapper[4954]: W1209 17:30:12.238905 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod507b400c_fcae_4257_a6e5_2485da532750.slice/crio-851aa647fc7d9646eb4f96017a9dc0e25994139d8f96a9121a32c64a4b43d762 WatchSource:0}: Error finding container 851aa647fc7d9646eb4f96017a9dc0e25994139d8f96a9121a32c64a4b43d762: Status 404 returned error can't find the container with id 851aa647fc7d9646eb4f96017a9dc0e25994139d8f96a9121a32c64a4b43d762 Dec 09 17:30:12 crc kubenswrapper[4954]: I1209 17:30:12.746054 4954 generic.go:334] "Generic (PLEG): container finished" podID="507b400c-fcae-4257-a6e5-2485da532750" containerID="e3fa3ea2965b62e667fd670ca8dcfedf5079b9846de4579248fab4a0b8660b7a" exitCode=0 Dec 09 17:30:12 crc kubenswrapper[4954]: I1209 17:30:12.746147 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" event={"ID":"507b400c-fcae-4257-a6e5-2485da532750","Type":"ContainerDied","Data":"e3fa3ea2965b62e667fd670ca8dcfedf5079b9846de4579248fab4a0b8660b7a"} Dec 09 17:30:12 crc kubenswrapper[4954]: I1209 17:30:12.747090 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" event={"ID":"507b400c-fcae-4257-a6e5-2485da532750","Type":"ContainerStarted","Data":"851aa647fc7d9646eb4f96017a9dc0e25994139d8f96a9121a32c64a4b43d762"} Dec 09 17:30:13 crc kubenswrapper[4954]: I1209 17:30:13.761992 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" event={"ID":"507b400c-fcae-4257-a6e5-2485da532750","Type":"ContainerStarted","Data":"13799642b80ec8b052de87ce60ecb3f5d5f44d5347698e50b625609de5897637"} Dec 09 17:30:13 crc kubenswrapper[4954]: I1209 17:30:13.762655 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:13 crc kubenswrapper[4954]: I1209 17:30:13.785692 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" podStartSLOduration=2.785667251 podStartE2EDuration="2.785667251s" podCreationTimestamp="2025-12-09 17:30:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:30:13.781529992 +0000 UTC m=+2010.169703822" watchObservedRunningTime="2025-12-09 17:30:13.785667251 +0000 UTC m=+2010.173841071" Dec 09 17:30:13 crc kubenswrapper[4954]: I1209 17:30:13.857559 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.361542 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.362201 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerName="nova-api-log" containerID="cri-o://49603484aa52c31a7f74ce4ecf682f4d2a0fc175479035cb29c2e14aa10e9cf5" gracePeriod=30 Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.362688 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerName="nova-api-api" containerID="cri-o://ead158534d1701a55c24f0c87725e83d27fddaa15c59187aebe22aa92e53e15e" gracePeriod=30 Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.776159 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="ceilometer-central-agent" containerID="cri-o://4e02788a4e2e3301d0f021f658a3d4525948d82ae4c64555a7ba09ef764c5e97" gracePeriod=30 Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.776222 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="sg-core" containerID="cri-o://3409e8d38448d1e74a39cc00ac58406a384efa8ebce20565943fb4118fa1f483" gracePeriod=30 Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.776171 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="proxy-httpd" containerID="cri-o://6fe1c2fb5cbd7762f197077949dfa7152bcc1e125d30f46c62f7a3d608da532d" gracePeriod=30 Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.776350 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c3d434-a778-48db-87d6-d96942b8578a","Type":"ContainerStarted","Data":"6fe1c2fb5cbd7762f197077949dfa7152bcc1e125d30f46c62f7a3d608da532d"} Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.776300 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="ceilometer-notification-agent" containerID="cri-o://3aaba52528c9b989ba9cfc2bcf2d55d164d167b06f23f7a58e435e19b0d98544" gracePeriod=30 Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.776717 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.780254 4954 generic.go:334] "Generic (PLEG): container finished" podID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerID="49603484aa52c31a7f74ce4ecf682f4d2a0fc175479035cb29c2e14aa10e9cf5" exitCode=143 Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.780346 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"af2878c3-2924-452b-93a9-0a2a843cfdff","Type":"ContainerDied","Data":"49603484aa52c31a7f74ce4ecf682f4d2a0fc175479035cb29c2e14aa10e9cf5"} Dec 09 17:30:14 crc kubenswrapper[4954]: I1209 17:30:14.811434 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.73694954 podStartE2EDuration="17.811410973s" podCreationTimestamp="2025-12-09 17:29:57 +0000 UTC" firstStartedPulling="2025-12-09 17:29:58.855568938 +0000 UTC m=+1995.243742758" lastFinishedPulling="2025-12-09 17:30:13.930030371 +0000 UTC m=+2010.318204191" observedRunningTime="2025-12-09 17:30:14.79950002 +0000 UTC m=+2011.187673850" watchObservedRunningTime="2025-12-09 17:30:14.811410973 +0000 UTC m=+2011.199584793" Dec 09 17:30:15 crc kubenswrapper[4954]: I1209 17:30:15.798274 4954 generic.go:334] "Generic (PLEG): container finished" podID="11c3d434-a778-48db-87d6-d96942b8578a" containerID="6fe1c2fb5cbd7762f197077949dfa7152bcc1e125d30f46c62f7a3d608da532d" exitCode=0 Dec 09 17:30:15 crc kubenswrapper[4954]: I1209 17:30:15.798679 4954 generic.go:334] "Generic (PLEG): container finished" podID="11c3d434-a778-48db-87d6-d96942b8578a" containerID="3409e8d38448d1e74a39cc00ac58406a384efa8ebce20565943fb4118fa1f483" exitCode=2 Dec 09 17:30:15 crc kubenswrapper[4954]: I1209 17:30:15.798695 4954 generic.go:334] "Generic (PLEG): container finished" podID="11c3d434-a778-48db-87d6-d96942b8578a" containerID="3aaba52528c9b989ba9cfc2bcf2d55d164d167b06f23f7a58e435e19b0d98544" exitCode=0 Dec 09 17:30:15 crc kubenswrapper[4954]: I1209 17:30:15.798706 4954 generic.go:334] "Generic (PLEG): container finished" podID="11c3d434-a778-48db-87d6-d96942b8578a" containerID="4e02788a4e2e3301d0f021f658a3d4525948d82ae4c64555a7ba09ef764c5e97" exitCode=0 Dec 09 17:30:15 crc kubenswrapper[4954]: I1209 17:30:15.798439 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c3d434-a778-48db-87d6-d96942b8578a","Type":"ContainerDied","Data":"6fe1c2fb5cbd7762f197077949dfa7152bcc1e125d30f46c62f7a3d608da532d"} Dec 09 17:30:15 crc kubenswrapper[4954]: I1209 17:30:15.798757 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c3d434-a778-48db-87d6-d96942b8578a","Type":"ContainerDied","Data":"3409e8d38448d1e74a39cc00ac58406a384efa8ebce20565943fb4118fa1f483"} Dec 09 17:30:15 crc kubenswrapper[4954]: I1209 17:30:15.798776 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c3d434-a778-48db-87d6-d96942b8578a","Type":"ContainerDied","Data":"3aaba52528c9b989ba9cfc2bcf2d55d164d167b06f23f7a58e435e19b0d98544"} Dec 09 17:30:15 crc kubenswrapper[4954]: I1209 17:30:15.798790 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c3d434-a778-48db-87d6-d96942b8578a","Type":"ContainerDied","Data":"4e02788a4e2e3301d0f021f658a3d4525948d82ae4c64555a7ba09ef764c5e97"} Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.122042 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.237765 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-run-httpd\") pod \"11c3d434-a778-48db-87d6-d96942b8578a\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.237860 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-combined-ca-bundle\") pod \"11c3d434-a778-48db-87d6-d96942b8578a\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.237959 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-sg-core-conf-yaml\") pod \"11c3d434-a778-48db-87d6-d96942b8578a\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.238041 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjvkt\" (UniqueName: \"kubernetes.io/projected/11c3d434-a778-48db-87d6-d96942b8578a-kube-api-access-fjvkt\") pod \"11c3d434-a778-48db-87d6-d96942b8578a\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.238095 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-scripts\") pod \"11c3d434-a778-48db-87d6-d96942b8578a\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.238242 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-config-data\") pod \"11c3d434-a778-48db-87d6-d96942b8578a\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.238327 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-log-httpd\") pod \"11c3d434-a778-48db-87d6-d96942b8578a\" (UID: \"11c3d434-a778-48db-87d6-d96942b8578a\") " Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.238409 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "11c3d434-a778-48db-87d6-d96942b8578a" (UID: "11c3d434-a778-48db-87d6-d96942b8578a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.238908 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "11c3d434-a778-48db-87d6-d96942b8578a" (UID: "11c3d434-a778-48db-87d6-d96942b8578a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.240361 4954 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.240385 4954 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11c3d434-a778-48db-87d6-d96942b8578a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.245950 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-scripts" (OuterVolumeSpecName: "scripts") pod "11c3d434-a778-48db-87d6-d96942b8578a" (UID: "11c3d434-a778-48db-87d6-d96942b8578a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.246430 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11c3d434-a778-48db-87d6-d96942b8578a-kube-api-access-fjvkt" (OuterVolumeSpecName: "kube-api-access-fjvkt") pod "11c3d434-a778-48db-87d6-d96942b8578a" (UID: "11c3d434-a778-48db-87d6-d96942b8578a"). InnerVolumeSpecName "kube-api-access-fjvkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.277403 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "11c3d434-a778-48db-87d6-d96942b8578a" (UID: "11c3d434-a778-48db-87d6-d96942b8578a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.325498 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11c3d434-a778-48db-87d6-d96942b8578a" (UID: "11c3d434-a778-48db-87d6-d96942b8578a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.347119 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.347300 4954 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.347371 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjvkt\" (UniqueName: \"kubernetes.io/projected/11c3d434-a778-48db-87d6-d96942b8578a-kube-api-access-fjvkt\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.347440 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.371440 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-config-data" (OuterVolumeSpecName: "config-data") pod "11c3d434-a778-48db-87d6-d96942b8578a" (UID: "11c3d434-a778-48db-87d6-d96942b8578a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.450669 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11c3d434-a778-48db-87d6-d96942b8578a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.818048 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11c3d434-a778-48db-87d6-d96942b8578a","Type":"ContainerDied","Data":"21e5095dacffc25c5ea7bb8bab538adb79138deb3a7692509709937bcbcbbe2c"} Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.818116 4954 scope.go:117] "RemoveContainer" containerID="6fe1c2fb5cbd7762f197077949dfa7152bcc1e125d30f46c62f7a3d608da532d" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.818160 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.865319 4954 scope.go:117] "RemoveContainer" containerID="3409e8d38448d1e74a39cc00ac58406a384efa8ebce20565943fb4118fa1f483" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.872950 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.890843 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.905958 4954 scope.go:117] "RemoveContainer" containerID="3aaba52528c9b989ba9cfc2bcf2d55d164d167b06f23f7a58e435e19b0d98544" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.909506 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:16 crc kubenswrapper[4954]: E1209 17:30:16.910135 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="sg-core" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.910161 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="sg-core" Dec 09 17:30:16 crc kubenswrapper[4954]: E1209 17:30:16.910188 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="ceilometer-central-agent" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.910198 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="ceilometer-central-agent" Dec 09 17:30:16 crc kubenswrapper[4954]: E1209 17:30:16.910231 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="ceilometer-notification-agent" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.910314 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="ceilometer-notification-agent" Dec 09 17:30:16 crc kubenswrapper[4954]: E1209 17:30:16.910337 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="proxy-httpd" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.910346 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="proxy-httpd" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.910643 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="proxy-httpd" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.910668 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="ceilometer-notification-agent" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.910679 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="sg-core" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.910691 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="11c3d434-a778-48db-87d6-d96942b8578a" containerName="ceilometer-central-agent" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.916260 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.920450 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.920640 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.930181 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:16 crc kubenswrapper[4954]: I1209 17:30:16.979034 4954 scope.go:117] "RemoveContainer" containerID="4e02788a4e2e3301d0f021f658a3d4525948d82ae4c64555a7ba09ef764c5e97" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.063741 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-run-httpd\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.063835 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.063894 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-log-httpd\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.063937 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-config-data\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.063965 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.064003 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwslc\" (UniqueName: \"kubernetes.io/projected/a1df5925-6ceb-466c-a3bd-0b3373c7351d-kube-api-access-zwslc\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.064069 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-scripts\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.101377 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.126805 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.166755 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-config-data\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.166824 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.166878 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwslc\" (UniqueName: \"kubernetes.io/projected/a1df5925-6ceb-466c-a3bd-0b3373c7351d-kube-api-access-zwslc\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.166968 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-scripts\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.167033 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-run-httpd\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.167103 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.167158 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-log-httpd\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.167807 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-log-httpd\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.167806 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-run-httpd\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.172201 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.174068 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-scripts\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.178915 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-config-data\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.181582 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.185554 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwslc\" (UniqueName: \"kubernetes.io/projected/a1df5925-6ceb-466c-a3bd-0b3373c7351d-kube-api-access-zwslc\") pod \"ceilometer-0\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.277714 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.777125 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.835435 4954 generic.go:334] "Generic (PLEG): container finished" podID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerID="ead158534d1701a55c24f0c87725e83d27fddaa15c59187aebe22aa92e53e15e" exitCode=0 Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.835504 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"af2878c3-2924-452b-93a9-0a2a843cfdff","Type":"ContainerDied","Data":"ead158534d1701a55c24f0c87725e83d27fddaa15c59187aebe22aa92e53e15e"} Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.841887 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1df5925-6ceb-466c-a3bd-0b3373c7351d","Type":"ContainerStarted","Data":"4120eb48ba18b7df26a4e7187992d7e0665c6cd463453be0036b3604ebc6093f"} Dec 09 17:30:17 crc kubenswrapper[4954]: I1209 17:30:17.867128 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.070655 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.105072 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-4bl8c"] Dec 09 17:30:18 crc kubenswrapper[4954]: E1209 17:30:18.105706 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerName="nova-api-log" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.105729 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerName="nova-api-log" Dec 09 17:30:18 crc kubenswrapper[4954]: E1209 17:30:18.105754 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerName="nova-api-api" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.105761 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerName="nova-api-api" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.105986 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerName="nova-api-log" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.106022 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" containerName="nova-api-api" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.106992 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.111796 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.112120 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.154570 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11c3d434-a778-48db-87d6-d96942b8578a" path="/var/lib/kubelet/pods/11c3d434-a778-48db-87d6-d96942b8578a/volumes" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.160880 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4bl8c"] Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.201273 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-combined-ca-bundle\") pod \"af2878c3-2924-452b-93a9-0a2a843cfdff\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.201389 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-config-data\") pod \"af2878c3-2924-452b-93a9-0a2a843cfdff\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.201576 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af2878c3-2924-452b-93a9-0a2a843cfdff-logs\") pod \"af2878c3-2924-452b-93a9-0a2a843cfdff\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.201691 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6prqn\" (UniqueName: \"kubernetes.io/projected/af2878c3-2924-452b-93a9-0a2a843cfdff-kube-api-access-6prqn\") pod \"af2878c3-2924-452b-93a9-0a2a843cfdff\" (UID: \"af2878c3-2924-452b-93a9-0a2a843cfdff\") " Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.202018 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.202118 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-scripts\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.202156 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqzj6\" (UniqueName: \"kubernetes.io/projected/961eb076-69ef-4160-80db-3b44f94f5a73-kube-api-access-hqzj6\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.202191 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-config-data\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.203390 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af2878c3-2924-452b-93a9-0a2a843cfdff-logs" (OuterVolumeSpecName: "logs") pod "af2878c3-2924-452b-93a9-0a2a843cfdff" (UID: "af2878c3-2924-452b-93a9-0a2a843cfdff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.228065 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af2878c3-2924-452b-93a9-0a2a843cfdff-kube-api-access-6prqn" (OuterVolumeSpecName: "kube-api-access-6prqn") pod "af2878c3-2924-452b-93a9-0a2a843cfdff" (UID: "af2878c3-2924-452b-93a9-0a2a843cfdff"). InnerVolumeSpecName "kube-api-access-6prqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.259587 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-config-data" (OuterVolumeSpecName: "config-data") pod "af2878c3-2924-452b-93a9-0a2a843cfdff" (UID: "af2878c3-2924-452b-93a9-0a2a843cfdff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.266085 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af2878c3-2924-452b-93a9-0a2a843cfdff" (UID: "af2878c3-2924-452b-93a9-0a2a843cfdff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.304642 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqzj6\" (UniqueName: \"kubernetes.io/projected/961eb076-69ef-4160-80db-3b44f94f5a73-kube-api-access-hqzj6\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.304735 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-config-data\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.304871 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.304972 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-scripts\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.305046 4954 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af2878c3-2924-452b-93a9-0a2a843cfdff-logs\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.305062 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6prqn\" (UniqueName: \"kubernetes.io/projected/af2878c3-2924-452b-93a9-0a2a843cfdff-kube-api-access-6prqn\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.305074 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.305083 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af2878c3-2924-452b-93a9-0a2a843cfdff-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.312193 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-scripts\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.312318 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-config-data\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.319760 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.331952 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqzj6\" (UniqueName: \"kubernetes.io/projected/961eb076-69ef-4160-80db-3b44f94f5a73-kube-api-access-hqzj6\") pod \"nova-cell1-cell-mapping-4bl8c\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.465540 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.874726 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"af2878c3-2924-452b-93a9-0a2a843cfdff","Type":"ContainerDied","Data":"474eaf2f7919a9c6cd02f00b06cc25b07099cc66f72edcb9453c487fc21e0eed"} Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.874769 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.874805 4954 scope.go:117] "RemoveContainer" containerID="ead158534d1701a55c24f0c87725e83d27fddaa15c59187aebe22aa92e53e15e" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.883062 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1df5925-6ceb-466c-a3bd-0b3373c7351d","Type":"ContainerStarted","Data":"d237c59ecd021445baf479358507e0aef7babd5f0631c0c91db2eda378dfbbdf"} Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.916877 4954 scope.go:117] "RemoveContainer" containerID="49603484aa52c31a7f74ce4ecf682f4d2a0fc175479035cb29c2e14aa10e9cf5" Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.940646 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:18 crc kubenswrapper[4954]: I1209 17:30:18.986956 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.002195 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.004715 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.012534 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.012996 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.013129 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.020112 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:19 crc kubenswrapper[4954]: W1209 17:30:19.037198 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod961eb076_69ef_4160_80db_3b44f94f5a73.slice/crio-7bc2213fee2c3512104c10827febac2dd6f3b0b0bcae988286c8d45c4b4d9a1a WatchSource:0}: Error finding container 7bc2213fee2c3512104c10827febac2dd6f3b0b0bcae988286c8d45c4b4d9a1a: Status 404 returned error can't find the container with id 7bc2213fee2c3512104c10827febac2dd6f3b0b0bcae988286c8d45c4b4d9a1a Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.047126 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4bl8c"] Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.139239 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.139366 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.139467 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-config-data\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.139568 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5548ae6f-a301-48e3-acbe-4ba125110356-logs\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.139818 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-public-tls-certs\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.140015 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25qn2\" (UniqueName: \"kubernetes.io/projected/5548ae6f-a301-48e3-acbe-4ba125110356-kube-api-access-25qn2\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.243526 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.243689 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-config-data\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.243725 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5548ae6f-a301-48e3-acbe-4ba125110356-logs\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.243792 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-public-tls-certs\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.243845 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25qn2\" (UniqueName: \"kubernetes.io/projected/5548ae6f-a301-48e3-acbe-4ba125110356-kube-api-access-25qn2\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.244147 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.247262 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5548ae6f-a301-48e3-acbe-4ba125110356-logs\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.248759 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-config-data\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.252062 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-public-tls-certs\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.252644 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.255311 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.274470 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25qn2\" (UniqueName: \"kubernetes.io/projected/5548ae6f-a301-48e3-acbe-4ba125110356-kube-api-access-25qn2\") pod \"nova-api-0\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.343692 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:30:19 crc kubenswrapper[4954]: W1209 17:30:19.825805 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5548ae6f_a301_48e3_acbe_4ba125110356.slice/crio-ff2897fc9c90cf5b7f8bd15a2743e55bf62e3b1912794469579e7b3216423426 WatchSource:0}: Error finding container ff2897fc9c90cf5b7f8bd15a2743e55bf62e3b1912794469579e7b3216423426: Status 404 returned error can't find the container with id ff2897fc9c90cf5b7f8bd15a2743e55bf62e3b1912794469579e7b3216423426 Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.828521 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.902972 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5548ae6f-a301-48e3-acbe-4ba125110356","Type":"ContainerStarted","Data":"ff2897fc9c90cf5b7f8bd15a2743e55bf62e3b1912794469579e7b3216423426"} Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.909800 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4bl8c" event={"ID":"961eb076-69ef-4160-80db-3b44f94f5a73","Type":"ContainerStarted","Data":"1060380ad794329e8565edb3dda2d6962aeb50e7069307bca0a9fc23e956bb26"} Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.910534 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4bl8c" event={"ID":"961eb076-69ef-4160-80db-3b44f94f5a73","Type":"ContainerStarted","Data":"7bc2213fee2c3512104c10827febac2dd6f3b0b0bcae988286c8d45c4b4d9a1a"} Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.914154 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1df5925-6ceb-466c-a3bd-0b3373c7351d","Type":"ContainerStarted","Data":"ff078a7a54cf55d551ed7b1cc84fd4dc3a7ee136d9797be4beb31a5c0b4022f6"} Dec 09 17:30:19 crc kubenswrapper[4954]: I1209 17:30:19.935191 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-4bl8c" podStartSLOduration=1.9351665439999999 podStartE2EDuration="1.935166544s" podCreationTimestamp="2025-12-09 17:30:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:30:19.93438045 +0000 UTC m=+2016.322554290" watchObservedRunningTime="2025-12-09 17:30:19.935166544 +0000 UTC m=+2016.323340364" Dec 09 17:30:20 crc kubenswrapper[4954]: I1209 17:30:20.149193 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af2878c3-2924-452b-93a9-0a2a843cfdff" path="/var/lib/kubelet/pods/af2878c3-2924-452b-93a9-0a2a843cfdff/volumes" Dec 09 17:30:20 crc kubenswrapper[4954]: I1209 17:30:20.932110 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5548ae6f-a301-48e3-acbe-4ba125110356","Type":"ContainerStarted","Data":"ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da"} Dec 09 17:30:20 crc kubenswrapper[4954]: I1209 17:30:20.932430 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5548ae6f-a301-48e3-acbe-4ba125110356","Type":"ContainerStarted","Data":"1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74"} Dec 09 17:30:20 crc kubenswrapper[4954]: I1209 17:30:20.935290 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1df5925-6ceb-466c-a3bd-0b3373c7351d","Type":"ContainerStarted","Data":"4be6620af5968a980115be19bd7c8ebe122982fe16e90eb212a05e310652bf55"} Dec 09 17:30:20 crc kubenswrapper[4954]: I1209 17:30:20.969827 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.969801194 podStartE2EDuration="2.969801194s" podCreationTimestamp="2025-12-09 17:30:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:30:20.965016124 +0000 UTC m=+2017.353189954" watchObservedRunningTime="2025-12-09 17:30:20.969801194 +0000 UTC m=+2017.357975034" Dec 09 17:30:21 crc kubenswrapper[4954]: I1209 17:30:21.728784 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:30:21 crc kubenswrapper[4954]: I1209 17:30:21.840369 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gbvxf"] Dec 09 17:30:21 crc kubenswrapper[4954]: I1209 17:30:21.840878 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" podUID="8198eada-d79f-4b68-877c-35325473c747" containerName="dnsmasq-dns" containerID="cri-o://bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c" gracePeriod=10 Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.464282 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.631544 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm9gw\" (UniqueName: \"kubernetes.io/projected/8198eada-d79f-4b68-877c-35325473c747-kube-api-access-nm9gw\") pod \"8198eada-d79f-4b68-877c-35325473c747\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.631803 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-sb\") pod \"8198eada-d79f-4b68-877c-35325473c747\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.631856 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-nb\") pod \"8198eada-d79f-4b68-877c-35325473c747\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.631889 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-svc\") pod \"8198eada-d79f-4b68-877c-35325473c747\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.632007 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-swift-storage-0\") pod \"8198eada-d79f-4b68-877c-35325473c747\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.632061 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-config\") pod \"8198eada-d79f-4b68-877c-35325473c747\" (UID: \"8198eada-d79f-4b68-877c-35325473c747\") " Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.642301 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8198eada-d79f-4b68-877c-35325473c747-kube-api-access-nm9gw" (OuterVolumeSpecName: "kube-api-access-nm9gw") pod "8198eada-d79f-4b68-877c-35325473c747" (UID: "8198eada-d79f-4b68-877c-35325473c747"). InnerVolumeSpecName "kube-api-access-nm9gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.699511 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-config" (OuterVolumeSpecName: "config") pod "8198eada-d79f-4b68-877c-35325473c747" (UID: "8198eada-d79f-4b68-877c-35325473c747"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.709089 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8198eada-d79f-4b68-877c-35325473c747" (UID: "8198eada-d79f-4b68-877c-35325473c747"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.715741 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8198eada-d79f-4b68-877c-35325473c747" (UID: "8198eada-d79f-4b68-877c-35325473c747"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.718244 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8198eada-d79f-4b68-877c-35325473c747" (UID: "8198eada-d79f-4b68-877c-35325473c747"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.720142 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8198eada-d79f-4b68-877c-35325473c747" (UID: "8198eada-d79f-4b68-877c-35325473c747"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.735232 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.735269 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.735285 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.735326 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.735337 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8198eada-d79f-4b68-877c-35325473c747-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.735346 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm9gw\" (UniqueName: \"kubernetes.io/projected/8198eada-d79f-4b68-877c-35325473c747-kube-api-access-nm9gw\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.970421 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1df5925-6ceb-466c-a3bd-0b3373c7351d","Type":"ContainerStarted","Data":"d9e65b648349a288fb93f22404ff0b315eeca9b04e2cd46e7428d85e95c4132d"} Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.970572 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.972501 4954 generic.go:334] "Generic (PLEG): container finished" podID="8198eada-d79f-4b68-877c-35325473c747" containerID="bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c" exitCode=0 Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.972542 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" event={"ID":"8198eada-d79f-4b68-877c-35325473c747","Type":"ContainerDied","Data":"bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c"} Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.972573 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" event={"ID":"8198eada-d79f-4b68-877c-35325473c747","Type":"ContainerDied","Data":"4e36f587cf39ff7f930d3ec1e509aebba3b5980055890d95de5eb610dbbe55ee"} Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.972607 4954 scope.go:117] "RemoveContainer" containerID="bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c" Dec 09 17:30:22 crc kubenswrapper[4954]: I1209 17:30:22.972611 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-gbvxf" Dec 09 17:30:23 crc kubenswrapper[4954]: I1209 17:30:23.021830 4954 scope.go:117] "RemoveContainer" containerID="732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454" Dec 09 17:30:23 crc kubenswrapper[4954]: I1209 17:30:23.026572 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.047364503 podStartE2EDuration="7.026543252s" podCreationTimestamp="2025-12-09 17:30:16 +0000 UTC" firstStartedPulling="2025-12-09 17:30:17.758039699 +0000 UTC m=+2014.146213519" lastFinishedPulling="2025-12-09 17:30:21.737218448 +0000 UTC m=+2018.125392268" observedRunningTime="2025-12-09 17:30:23.016501027 +0000 UTC m=+2019.404674857" watchObservedRunningTime="2025-12-09 17:30:23.026543252 +0000 UTC m=+2019.414717072" Dec 09 17:30:23 crc kubenswrapper[4954]: I1209 17:30:23.077659 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gbvxf"] Dec 09 17:30:23 crc kubenswrapper[4954]: I1209 17:30:23.097350 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-gbvxf"] Dec 09 17:30:23 crc kubenswrapper[4954]: I1209 17:30:23.098857 4954 scope.go:117] "RemoveContainer" containerID="bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c" Dec 09 17:30:23 crc kubenswrapper[4954]: E1209 17:30:23.104828 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c\": container with ID starting with bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c not found: ID does not exist" containerID="bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c" Dec 09 17:30:23 crc kubenswrapper[4954]: I1209 17:30:23.105153 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c"} err="failed to get container status \"bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c\": rpc error: code = NotFound desc = could not find container \"bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c\": container with ID starting with bfd4a904e3389a975e6be62fb8f8596e6e5829265f2c7292cd45613839312e9c not found: ID does not exist" Dec 09 17:30:23 crc kubenswrapper[4954]: I1209 17:30:23.105199 4954 scope.go:117] "RemoveContainer" containerID="732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454" Dec 09 17:30:23 crc kubenswrapper[4954]: E1209 17:30:23.111796 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454\": container with ID starting with 732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454 not found: ID does not exist" containerID="732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454" Dec 09 17:30:23 crc kubenswrapper[4954]: I1209 17:30:23.111851 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454"} err="failed to get container status \"732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454\": rpc error: code = NotFound desc = could not find container \"732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454\": container with ID starting with 732272dc3964fa6ea669abeefc47d14795a9b11cf9d0677b7e8fb0175147e454 not found: ID does not exist" Dec 09 17:30:24 crc kubenswrapper[4954]: I1209 17:30:24.135341 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8198eada-d79f-4b68-877c-35325473c747" path="/var/lib/kubelet/pods/8198eada-d79f-4b68-877c-35325473c747/volumes" Dec 09 17:30:26 crc kubenswrapper[4954]: I1209 17:30:26.015909 4954 generic.go:334] "Generic (PLEG): container finished" podID="961eb076-69ef-4160-80db-3b44f94f5a73" containerID="1060380ad794329e8565edb3dda2d6962aeb50e7069307bca0a9fc23e956bb26" exitCode=0 Dec 09 17:30:26 crc kubenswrapper[4954]: I1209 17:30:26.016047 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4bl8c" event={"ID":"961eb076-69ef-4160-80db-3b44f94f5a73","Type":"ContainerDied","Data":"1060380ad794329e8565edb3dda2d6962aeb50e7069307bca0a9fc23e956bb26"} Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.436210 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.559012 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-combined-ca-bundle\") pod \"961eb076-69ef-4160-80db-3b44f94f5a73\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.559114 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqzj6\" (UniqueName: \"kubernetes.io/projected/961eb076-69ef-4160-80db-3b44f94f5a73-kube-api-access-hqzj6\") pod \"961eb076-69ef-4160-80db-3b44f94f5a73\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.559246 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-scripts\") pod \"961eb076-69ef-4160-80db-3b44f94f5a73\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.559365 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-config-data\") pod \"961eb076-69ef-4160-80db-3b44f94f5a73\" (UID: \"961eb076-69ef-4160-80db-3b44f94f5a73\") " Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.574455 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-scripts" (OuterVolumeSpecName: "scripts") pod "961eb076-69ef-4160-80db-3b44f94f5a73" (UID: "961eb076-69ef-4160-80db-3b44f94f5a73"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.575769 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/961eb076-69ef-4160-80db-3b44f94f5a73-kube-api-access-hqzj6" (OuterVolumeSpecName: "kube-api-access-hqzj6") pod "961eb076-69ef-4160-80db-3b44f94f5a73" (UID: "961eb076-69ef-4160-80db-3b44f94f5a73"). InnerVolumeSpecName "kube-api-access-hqzj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.594132 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "961eb076-69ef-4160-80db-3b44f94f5a73" (UID: "961eb076-69ef-4160-80db-3b44f94f5a73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.597762 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-config-data" (OuterVolumeSpecName: "config-data") pod "961eb076-69ef-4160-80db-3b44f94f5a73" (UID: "961eb076-69ef-4160-80db-3b44f94f5a73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.662091 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.662132 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqzj6\" (UniqueName: \"kubernetes.io/projected/961eb076-69ef-4160-80db-3b44f94f5a73-kube-api-access-hqzj6\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.662146 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:27 crc kubenswrapper[4954]: I1209 17:30:27.662159 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/961eb076-69ef-4160-80db-3b44f94f5a73-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.040505 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4bl8c" event={"ID":"961eb076-69ef-4160-80db-3b44f94f5a73","Type":"ContainerDied","Data":"7bc2213fee2c3512104c10827febac2dd6f3b0b0bcae988286c8d45c4b4d9a1a"} Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.040562 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bc2213fee2c3512104c10827febac2dd6f3b0b0bcae988286c8d45c4b4d9a1a" Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.040608 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4bl8c" Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.243244 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.243749 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5548ae6f-a301-48e3-acbe-4ba125110356" containerName="nova-api-log" containerID="cri-o://1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74" gracePeriod=30 Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.243886 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5548ae6f-a301-48e3-acbe-4ba125110356" containerName="nova-api-api" containerID="cri-o://ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da" gracePeriod=30 Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.278383 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.280248 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c2b4d601-8a86-48b0-abd3-039526cb6057" containerName="nova-scheduler-scheduler" containerID="cri-o://fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1" gracePeriod=30 Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.303889 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.304163 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-log" containerID="cri-o://08f34ee49683df1db9d34a0f255a676999ad9be2ba72e974eec23eb81c93a0ba" gracePeriod=30 Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.304703 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-metadata" containerID="cri-o://eb6a4c99ea7a347b0d2674943908b4ff21ab0ee5c6078de421b64f30dfcc7309" gracePeriod=30 Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.902108 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.992582 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-combined-ca-bundle\") pod \"5548ae6f-a301-48e3-acbe-4ba125110356\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.992721 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5548ae6f-a301-48e3-acbe-4ba125110356-logs\") pod \"5548ae6f-a301-48e3-acbe-4ba125110356\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.992914 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25qn2\" (UniqueName: \"kubernetes.io/projected/5548ae6f-a301-48e3-acbe-4ba125110356-kube-api-access-25qn2\") pod \"5548ae6f-a301-48e3-acbe-4ba125110356\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.993093 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-internal-tls-certs\") pod \"5548ae6f-a301-48e3-acbe-4ba125110356\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.993121 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-config-data\") pod \"5548ae6f-a301-48e3-acbe-4ba125110356\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.993167 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-public-tls-certs\") pod \"5548ae6f-a301-48e3-acbe-4ba125110356\" (UID: \"5548ae6f-a301-48e3-acbe-4ba125110356\") " Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.994184 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5548ae6f-a301-48e3-acbe-4ba125110356-logs" (OuterVolumeSpecName: "logs") pod "5548ae6f-a301-48e3-acbe-4ba125110356" (UID: "5548ae6f-a301-48e3-acbe-4ba125110356"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:30:28 crc kubenswrapper[4954]: I1209 17:30:28.999868 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5548ae6f-a301-48e3-acbe-4ba125110356-kube-api-access-25qn2" (OuterVolumeSpecName: "kube-api-access-25qn2") pod "5548ae6f-a301-48e3-acbe-4ba125110356" (UID: "5548ae6f-a301-48e3-acbe-4ba125110356"). InnerVolumeSpecName "kube-api-access-25qn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.028287 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-config-data" (OuterVolumeSpecName: "config-data") pod "5548ae6f-a301-48e3-acbe-4ba125110356" (UID: "5548ae6f-a301-48e3-acbe-4ba125110356"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.028664 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5548ae6f-a301-48e3-acbe-4ba125110356" (UID: "5548ae6f-a301-48e3-acbe-4ba125110356"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.053421 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5548ae6f-a301-48e3-acbe-4ba125110356" (UID: "5548ae6f-a301-48e3-acbe-4ba125110356"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.054366 4954 generic.go:334] "Generic (PLEG): container finished" podID="5548ae6f-a301-48e3-acbe-4ba125110356" containerID="ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da" exitCode=0 Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.054406 4954 generic.go:334] "Generic (PLEG): container finished" podID="5548ae6f-a301-48e3-acbe-4ba125110356" containerID="1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74" exitCode=143 Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.054411 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5548ae6f-a301-48e3-acbe-4ba125110356","Type":"ContainerDied","Data":"ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da"} Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.054440 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.054450 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5548ae6f-a301-48e3-acbe-4ba125110356","Type":"ContainerDied","Data":"1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74"} Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.054462 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5548ae6f-a301-48e3-acbe-4ba125110356","Type":"ContainerDied","Data":"ff2897fc9c90cf5b7f8bd15a2743e55bf62e3b1912794469579e7b3216423426"} Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.054478 4954 scope.go:117] "RemoveContainer" containerID="ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.057855 4954 generic.go:334] "Generic (PLEG): container finished" podID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerID="08f34ee49683df1db9d34a0f255a676999ad9be2ba72e974eec23eb81c93a0ba" exitCode=143 Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.057901 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c","Type":"ContainerDied","Data":"08f34ee49683df1db9d34a0f255a676999ad9be2ba72e974eec23eb81c93a0ba"} Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.068888 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5548ae6f-a301-48e3-acbe-4ba125110356" (UID: "5548ae6f-a301-48e3-acbe-4ba125110356"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.084251 4954 scope.go:117] "RemoveContainer" containerID="1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.096875 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.096915 4954 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5548ae6f-a301-48e3-acbe-4ba125110356-logs\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.096925 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25qn2\" (UniqueName: \"kubernetes.io/projected/5548ae6f-a301-48e3-acbe-4ba125110356-kube-api-access-25qn2\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.096939 4954 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.096948 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.096959 4954 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5548ae6f-a301-48e3-acbe-4ba125110356-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.108392 4954 scope.go:117] "RemoveContainer" containerID="ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da" Dec 09 17:30:29 crc kubenswrapper[4954]: E1209 17:30:29.109155 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da\": container with ID starting with ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da not found: ID does not exist" containerID="ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.109190 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da"} err="failed to get container status \"ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da\": rpc error: code = NotFound desc = could not find container \"ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da\": container with ID starting with ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da not found: ID does not exist" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.109212 4954 scope.go:117] "RemoveContainer" containerID="1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74" Dec 09 17:30:29 crc kubenswrapper[4954]: E1209 17:30:29.109690 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74\": container with ID starting with 1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74 not found: ID does not exist" containerID="1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.109730 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74"} err="failed to get container status \"1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74\": rpc error: code = NotFound desc = could not find container \"1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74\": container with ID starting with 1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74 not found: ID does not exist" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.109743 4954 scope.go:117] "RemoveContainer" containerID="ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.111899 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da"} err="failed to get container status \"ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da\": rpc error: code = NotFound desc = could not find container \"ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da\": container with ID starting with ac2ac2454a11cd4a1b5638afcd1c5f6ce04e847040d6f142270542dd8d7be9da not found: ID does not exist" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.111936 4954 scope.go:117] "RemoveContainer" containerID="1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.112210 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74"} err="failed to get container status \"1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74\": rpc error: code = NotFound desc = could not find container \"1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74\": container with ID starting with 1e869bbe07818d25a052c34a665b0ea7cbe0db714005be69bef612f1190cee74 not found: ID does not exist" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.442143 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.454851 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.468615 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:29 crc kubenswrapper[4954]: E1209 17:30:29.470283 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5548ae6f-a301-48e3-acbe-4ba125110356" containerName="nova-api-log" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.470334 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5548ae6f-a301-48e3-acbe-4ba125110356" containerName="nova-api-log" Dec 09 17:30:29 crc kubenswrapper[4954]: E1209 17:30:29.470374 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8198eada-d79f-4b68-877c-35325473c747" containerName="dnsmasq-dns" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.470383 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="8198eada-d79f-4b68-877c-35325473c747" containerName="dnsmasq-dns" Dec 09 17:30:29 crc kubenswrapper[4954]: E1209 17:30:29.470422 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5548ae6f-a301-48e3-acbe-4ba125110356" containerName="nova-api-api" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.470429 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5548ae6f-a301-48e3-acbe-4ba125110356" containerName="nova-api-api" Dec 09 17:30:29 crc kubenswrapper[4954]: E1209 17:30:29.470451 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8198eada-d79f-4b68-877c-35325473c747" containerName="init" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.470461 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="8198eada-d79f-4b68-877c-35325473c747" containerName="init" Dec 09 17:30:29 crc kubenswrapper[4954]: E1209 17:30:29.470479 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="961eb076-69ef-4160-80db-3b44f94f5a73" containerName="nova-manage" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.470485 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="961eb076-69ef-4160-80db-3b44f94f5a73" containerName="nova-manage" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.470886 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5548ae6f-a301-48e3-acbe-4ba125110356" containerName="nova-api-api" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.470914 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="8198eada-d79f-4b68-877c-35325473c747" containerName="dnsmasq-dns" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.470935 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5548ae6f-a301-48e3-acbe-4ba125110356" containerName="nova-api-log" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.470950 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="961eb076-69ef-4160-80db-3b44f94f5a73" containerName="nova-manage" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.472639 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.475189 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.475697 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.475966 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.489900 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.608149 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01db8384-81a2-4c9d-b26d-a8d788614768-logs\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.608197 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-public-tls-certs\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.608512 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2dcm\" (UniqueName: \"kubernetes.io/projected/01db8384-81a2-4c9d-b26d-a8d788614768-kube-api-access-j2dcm\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.608747 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-config-data\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.608871 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.608901 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-internal-tls-certs\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.711582 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-public-tls-certs\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.711720 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2dcm\" (UniqueName: \"kubernetes.io/projected/01db8384-81a2-4c9d-b26d-a8d788614768-kube-api-access-j2dcm\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.711771 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-config-data\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.711842 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.711864 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-internal-tls-certs\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.711972 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01db8384-81a2-4c9d-b26d-a8d788614768-logs\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.712481 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01db8384-81a2-4c9d-b26d-a8d788614768-logs\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.715842 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.716004 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-config-data\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.718162 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-public-tls-certs\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.719001 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01db8384-81a2-4c9d-b26d-a8d788614768-internal-tls-certs\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.728488 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2dcm\" (UniqueName: \"kubernetes.io/projected/01db8384-81a2-4c9d-b26d-a8d788614768-kube-api-access-j2dcm\") pod \"nova-api-0\" (UID: \"01db8384-81a2-4c9d-b26d-a8d788614768\") " pod="openstack/nova-api-0" Dec 09 17:30:29 crc kubenswrapper[4954]: I1209 17:30:29.793526 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 17:30:30 crc kubenswrapper[4954]: I1209 17:30:30.134733 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5548ae6f-a301-48e3-acbe-4ba125110356" path="/var/lib/kubelet/pods/5548ae6f-a301-48e3-acbe-4ba125110356/volumes" Dec 09 17:30:30 crc kubenswrapper[4954]: I1209 17:30:30.290299 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 17:30:30 crc kubenswrapper[4954]: E1209 17:30:30.836609 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1 is running failed: container process not found" containerID="fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 17:30:30 crc kubenswrapper[4954]: E1209 17:30:30.837502 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1 is running failed: container process not found" containerID="fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 17:30:30 crc kubenswrapper[4954]: E1209 17:30:30.837846 4954 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1 is running failed: container process not found" containerID="fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 17:30:30 crc kubenswrapper[4954]: E1209 17:30:30.837928 4954 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c2b4d601-8a86-48b0-abd3-039526cb6057" containerName="nova-scheduler-scheduler" Dec 09 17:30:30 crc kubenswrapper[4954]: I1209 17:30:30.847905 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 17:30:30 crc kubenswrapper[4954]: I1209 17:30:30.941713 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lczf\" (UniqueName: \"kubernetes.io/projected/c2b4d601-8a86-48b0-abd3-039526cb6057-kube-api-access-7lczf\") pod \"c2b4d601-8a86-48b0-abd3-039526cb6057\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " Dec 09 17:30:30 crc kubenswrapper[4954]: I1209 17:30:30.941783 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-config-data\") pod \"c2b4d601-8a86-48b0-abd3-039526cb6057\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " Dec 09 17:30:30 crc kubenswrapper[4954]: I1209 17:30:30.941928 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-combined-ca-bundle\") pod \"c2b4d601-8a86-48b0-abd3-039526cb6057\" (UID: \"c2b4d601-8a86-48b0-abd3-039526cb6057\") " Dec 09 17:30:30 crc kubenswrapper[4954]: I1209 17:30:30.961010 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b4d601-8a86-48b0-abd3-039526cb6057-kube-api-access-7lczf" (OuterVolumeSpecName: "kube-api-access-7lczf") pod "c2b4d601-8a86-48b0-abd3-039526cb6057" (UID: "c2b4d601-8a86-48b0-abd3-039526cb6057"). InnerVolumeSpecName "kube-api-access-7lczf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.002509 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-config-data" (OuterVolumeSpecName: "config-data") pod "c2b4d601-8a86-48b0-abd3-039526cb6057" (UID: "c2b4d601-8a86-48b0-abd3-039526cb6057"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.003734 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2b4d601-8a86-48b0-abd3-039526cb6057" (UID: "c2b4d601-8a86-48b0-abd3-039526cb6057"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.044341 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lczf\" (UniqueName: \"kubernetes.io/projected/c2b4d601-8a86-48b0-abd3-039526cb6057-kube-api-access-7lczf\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.044390 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.044405 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b4d601-8a86-48b0-abd3-039526cb6057-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.087282 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"01db8384-81a2-4c9d-b26d-a8d788614768","Type":"ContainerStarted","Data":"b758fc7a9fb96b4b98828bd7a70c59d3391c34c49108bdbd92becda07553f27d"} Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.087337 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"01db8384-81a2-4c9d-b26d-a8d788614768","Type":"ContainerStarted","Data":"3f01242be70dc8c7ab6966eca84591e3a118ca36d255d67d1c95c81fbaaa5604"} Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.087348 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"01db8384-81a2-4c9d-b26d-a8d788614768","Type":"ContainerStarted","Data":"7026ca10a49d36b6695c0cf62b8193b5a22bef7a9588d9e400a48a246c1a48d3"} Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.088693 4954 generic.go:334] "Generic (PLEG): container finished" podID="c2b4d601-8a86-48b0-abd3-039526cb6057" containerID="fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1" exitCode=0 Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.088733 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c2b4d601-8a86-48b0-abd3-039526cb6057","Type":"ContainerDied","Data":"fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1"} Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.088753 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c2b4d601-8a86-48b0-abd3-039526cb6057","Type":"ContainerDied","Data":"2aec6d21390d884799e01e04b960cdc11a2eb785e727527682acebe04e80647c"} Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.088773 4954 scope.go:117] "RemoveContainer" containerID="fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.088911 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.114639 4954 scope.go:117] "RemoveContainer" containerID="fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1" Dec 09 17:30:31 crc kubenswrapper[4954]: E1209 17:30:31.115275 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1\": container with ID starting with fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1 not found: ID does not exist" containerID="fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.115317 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1"} err="failed to get container status \"fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1\": rpc error: code = NotFound desc = could not find container \"fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1\": container with ID starting with fc0928ef05495f454e56ba0f87f848d4303e4ea21144dd72198165d672e403c1 not found: ID does not exist" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.122173 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.122152338 podStartE2EDuration="2.122152338s" podCreationTimestamp="2025-12-09 17:30:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:30:31.104440114 +0000 UTC m=+2027.492613944" watchObservedRunningTime="2025-12-09 17:30:31.122152338 +0000 UTC m=+2027.510326158" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.136831 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.157281 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.169319 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:30:31 crc kubenswrapper[4954]: E1209 17:30:31.169981 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b4d601-8a86-48b0-abd3-039526cb6057" containerName="nova-scheduler-scheduler" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.170003 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b4d601-8a86-48b0-abd3-039526cb6057" containerName="nova-scheduler-scheduler" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.170243 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2b4d601-8a86-48b0-abd3-039526cb6057" containerName="nova-scheduler-scheduler" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.171168 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.173768 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.180259 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.249220 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb9e244-80bc-4b7b-8b89-c44bdb73ae80-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80\") " pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.249302 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjfk8\" (UniqueName: \"kubernetes.io/projected/7cb9e244-80bc-4b7b-8b89-c44bdb73ae80-kube-api-access-mjfk8\") pod \"nova-scheduler-0\" (UID: \"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80\") " pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.249759 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb9e244-80bc-4b7b-8b89-c44bdb73ae80-config-data\") pod \"nova-scheduler-0\" (UID: \"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80\") " pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.352517 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb9e244-80bc-4b7b-8b89-c44bdb73ae80-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80\") " pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.352584 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjfk8\" (UniqueName: \"kubernetes.io/projected/7cb9e244-80bc-4b7b-8b89-c44bdb73ae80-kube-api-access-mjfk8\") pod \"nova-scheduler-0\" (UID: \"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80\") " pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.352884 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb9e244-80bc-4b7b-8b89-c44bdb73ae80-config-data\") pod \"nova-scheduler-0\" (UID: \"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80\") " pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.357321 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb9e244-80bc-4b7b-8b89-c44bdb73ae80-config-data\") pod \"nova-scheduler-0\" (UID: \"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80\") " pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.357873 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb9e244-80bc-4b7b-8b89-c44bdb73ae80-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80\") " pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.373054 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjfk8\" (UniqueName: \"kubernetes.io/projected/7cb9e244-80bc-4b7b-8b89-c44bdb73ae80-kube-api-access-mjfk8\") pod \"nova-scheduler-0\" (UID: \"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80\") " pod="openstack/nova-scheduler-0" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.465720 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.233:8775/\": read tcp 10.217.0.2:36534->10.217.0.233:8775: read: connection reset by peer" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.465720 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.233:8775/\": read tcp 10.217.0.2:36542->10.217.0.233:8775: read: connection reset by peer" Dec 09 17:30:31 crc kubenswrapper[4954]: I1209 17:30:31.493887 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.029575 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 17:30:32 crc kubenswrapper[4954]: W1209 17:30:32.033800 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cb9e244_80bc_4b7b_8b89_c44bdb73ae80.slice/crio-c57ee09bb725c82822fbe4c9bc8b6503191533f5d4c48b199ea5b3a450047c33 WatchSource:0}: Error finding container c57ee09bb725c82822fbe4c9bc8b6503191533f5d4c48b199ea5b3a450047c33: Status 404 returned error can't find the container with id c57ee09bb725c82822fbe4c9bc8b6503191533f5d4c48b199ea5b3a450047c33 Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.106536 4954 generic.go:334] "Generic (PLEG): container finished" podID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerID="eb6a4c99ea7a347b0d2674943908b4ff21ab0ee5c6078de421b64f30dfcc7309" exitCode=0 Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.106685 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c","Type":"ContainerDied","Data":"eb6a4c99ea7a347b0d2674943908b4ff21ab0ee5c6078de421b64f30dfcc7309"} Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.108286 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80","Type":"ContainerStarted","Data":"c57ee09bb725c82822fbe4c9bc8b6503191533f5d4c48b199ea5b3a450047c33"} Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.144807 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2b4d601-8a86-48b0-abd3-039526cb6057" path="/var/lib/kubelet/pods/c2b4d601-8a86-48b0-abd3-039526cb6057/volumes" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.477956 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.604560 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmfnj\" (UniqueName: \"kubernetes.io/projected/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-kube-api-access-rmfnj\") pod \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.604694 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-nova-metadata-tls-certs\") pod \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.605476 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-logs\") pod \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.605558 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-config-data\") pod \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.605774 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-combined-ca-bundle\") pod \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\" (UID: \"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c\") " Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.606680 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-logs" (OuterVolumeSpecName: "logs") pod "f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" (UID: "f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.607225 4954 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-logs\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.628729 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-kube-api-access-rmfnj" (OuterVolumeSpecName: "kube-api-access-rmfnj") pod "f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" (UID: "f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c"). InnerVolumeSpecName "kube-api-access-rmfnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.640774 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-config-data" (OuterVolumeSpecName: "config-data") pod "f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" (UID: "f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.652710 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" (UID: "f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.684569 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" (UID: "f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.709416 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.709454 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmfnj\" (UniqueName: \"kubernetes.io/projected/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-kube-api-access-rmfnj\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.709470 4954 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:32 crc kubenswrapper[4954]: I1209 17:30:32.709486 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.125094 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.126305 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c","Type":"ContainerDied","Data":"263edc65ea2f453fe2957d20215351179a3a24b84e6bdeb955dfd0adafb0de2a"} Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.126368 4954 scope.go:117] "RemoveContainer" containerID="eb6a4c99ea7a347b0d2674943908b4ff21ab0ee5c6078de421b64f30dfcc7309" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.126498 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.139271 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7cb9e244-80bc-4b7b-8b89-c44bdb73ae80","Type":"ContainerStarted","Data":"39603c5e1f7e94b74025e1e7cf3833b2be7b990e8542ad49b872b94c880d82f4"} Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.184330 4954 generic.go:334] "Generic (PLEG): container finished" podID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerID="ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491" exitCode=137 Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.184435 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"93f122ae-71b4-4393-a163-9f61a9b52b8d","Type":"ContainerDied","Data":"ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491"} Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.184480 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"93f122ae-71b4-4393-a163-9f61a9b52b8d","Type":"ContainerDied","Data":"0db99a42c5a92ef8cc6d078530dff14e39a64e18f80f92f595498a8de558c5f8"} Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.184655 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.200684 4954 scope.go:117] "RemoveContainer" containerID="08f34ee49683df1db9d34a0f255a676999ad9be2ba72e974eec23eb81c93a0ba" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.220226 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-combined-ca-bundle\") pod \"93f122ae-71b4-4393-a163-9f61a9b52b8d\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.220340 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xxj2\" (UniqueName: \"kubernetes.io/projected/93f122ae-71b4-4393-a163-9f61a9b52b8d-kube-api-access-4xxj2\") pod \"93f122ae-71b4-4393-a163-9f61a9b52b8d\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.220374 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-config-data\") pod \"93f122ae-71b4-4393-a163-9f61a9b52b8d\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.220457 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-scripts\") pod \"93f122ae-71b4-4393-a163-9f61a9b52b8d\" (UID: \"93f122ae-71b4-4393-a163-9f61a9b52b8d\") " Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.226089 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.226056722 podStartE2EDuration="2.226056722s" podCreationTimestamp="2025-12-09 17:30:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:30:33.192759499 +0000 UTC m=+2029.580933319" watchObservedRunningTime="2025-12-09 17:30:33.226056722 +0000 UTC m=+2029.614230542" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.246302 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-scripts" (OuterVolumeSpecName: "scripts") pod "93f122ae-71b4-4393-a163-9f61a9b52b8d" (UID: "93f122ae-71b4-4393-a163-9f61a9b52b8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.255865 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93f122ae-71b4-4393-a163-9f61a9b52b8d-kube-api-access-4xxj2" (OuterVolumeSpecName: "kube-api-access-4xxj2") pod "93f122ae-71b4-4393-a163-9f61a9b52b8d" (UID: "93f122ae-71b4-4393-a163-9f61a9b52b8d"). InnerVolumeSpecName "kube-api-access-4xxj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.269991 4954 scope.go:117] "RemoveContainer" containerID="ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.310329 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.322191 4954 scope.go:117] "RemoveContainer" containerID="3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.325621 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xxj2\" (UniqueName: \"kubernetes.io/projected/93f122ae-71b4-4393-a163-9f61a9b52b8d-kube-api-access-4xxj2\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.325664 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.337309 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.358820 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:30:33 crc kubenswrapper[4954]: E1209 17:30:33.359810 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-log" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.359840 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-log" Dec 09 17:30:33 crc kubenswrapper[4954]: E1209 17:30:33.359854 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-api" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.359862 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-api" Dec 09 17:30:33 crc kubenswrapper[4954]: E1209 17:30:33.359880 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-evaluator" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.359892 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-evaluator" Dec 09 17:30:33 crc kubenswrapper[4954]: E1209 17:30:33.359927 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-listener" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.359936 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-listener" Dec 09 17:30:33 crc kubenswrapper[4954]: E1209 17:30:33.359969 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-notifier" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.359976 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-notifier" Dec 09 17:30:33 crc kubenswrapper[4954]: E1209 17:30:33.359988 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-metadata" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.359995 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-metadata" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.360339 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-metadata" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.360373 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-api" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.360391 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-notifier" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.360404 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-evaluator" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.360419 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" containerName="nova-metadata-log" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.360430 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" containerName="aodh-listener" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.363062 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.376258 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.376990 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.381616 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.394020 4954 scope.go:117] "RemoveContainer" containerID="fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.422339 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93f122ae-71b4-4393-a163-9f61a9b52b8d" (UID: "93f122ae-71b4-4393-a163-9f61a9b52b8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.428830 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f299ddad-e1de-4ea5-a76c-35477739f14e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.428992 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f299ddad-e1de-4ea5-a76c-35477739f14e-config-data\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.429090 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f299ddad-e1de-4ea5-a76c-35477739f14e-logs\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.429208 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-589sf\" (UniqueName: \"kubernetes.io/projected/f299ddad-e1de-4ea5-a76c-35477739f14e-kube-api-access-589sf\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.429274 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f299ddad-e1de-4ea5-a76c-35477739f14e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.430583 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.434025 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-config-data" (OuterVolumeSpecName: "config-data") pod "93f122ae-71b4-4393-a163-9f61a9b52b8d" (UID: "93f122ae-71b4-4393-a163-9f61a9b52b8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.512735 4954 scope.go:117] "RemoveContainer" containerID="d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.532658 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f299ddad-e1de-4ea5-a76c-35477739f14e-logs\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.532770 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-589sf\" (UniqueName: \"kubernetes.io/projected/f299ddad-e1de-4ea5-a76c-35477739f14e-kube-api-access-589sf\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.532804 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f299ddad-e1de-4ea5-a76c-35477739f14e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.532909 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f299ddad-e1de-4ea5-a76c-35477739f14e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.532935 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f299ddad-e1de-4ea5-a76c-35477739f14e-config-data\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.533011 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f122ae-71b4-4393-a163-9f61a9b52b8d-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.534325 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f299ddad-e1de-4ea5-a76c-35477739f14e-logs\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.539484 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f299ddad-e1de-4ea5-a76c-35477739f14e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.539656 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f299ddad-e1de-4ea5-a76c-35477739f14e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.541698 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f299ddad-e1de-4ea5-a76c-35477739f14e-config-data\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.549676 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.557151 4954 scope.go:117] "RemoveContainer" containerID="ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491" Dec 09 17:30:33 crc kubenswrapper[4954]: E1209 17:30:33.557886 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491\": container with ID starting with ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491 not found: ID does not exist" containerID="ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.557928 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491"} err="failed to get container status \"ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491\": rpc error: code = NotFound desc = could not find container \"ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491\": container with ID starting with ab7953f092f02ac72b20bbc7b1eba3c11a5b05e8a96a3e87a4c18c176c5cd491 not found: ID does not exist" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.557952 4954 scope.go:117] "RemoveContainer" containerID="3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e" Dec 09 17:30:33 crc kubenswrapper[4954]: E1209 17:30:33.558422 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e\": container with ID starting with 3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e not found: ID does not exist" containerID="3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.558442 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e"} err="failed to get container status \"3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e\": rpc error: code = NotFound desc = could not find container \"3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e\": container with ID starting with 3834e476e3e3a71442eb53c1eb633b3c63589652db81acab633a5ebd2567536e not found: ID does not exist" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.558458 4954 scope.go:117] "RemoveContainer" containerID="fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f" Dec 09 17:30:33 crc kubenswrapper[4954]: E1209 17:30:33.561178 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f\": container with ID starting with fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f not found: ID does not exist" containerID="fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.561242 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f"} err="failed to get container status \"fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f\": rpc error: code = NotFound desc = could not find container \"fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f\": container with ID starting with fedbfedc81fffd01871662778aa70cd46c22d7ebaac6b38317b440eaa4ea545f not found: ID does not exist" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.561275 4954 scope.go:117] "RemoveContainer" containerID="d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.563672 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.565498 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-589sf\" (UniqueName: \"kubernetes.io/projected/f299ddad-e1de-4ea5-a76c-35477739f14e-kube-api-access-589sf\") pod \"nova-metadata-0\" (UID: \"f299ddad-e1de-4ea5-a76c-35477739f14e\") " pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: E1209 17:30:33.569808 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8\": container with ID starting with d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8 not found: ID does not exist" containerID="d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.569864 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8"} err="failed to get container status \"d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8\": rpc error: code = NotFound desc = could not find container \"d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8\": container with ID starting with d9afba6d271a5980796b9bb0ed4ebab35c007fa45a067b14d97536a91214b6b8 not found: ID does not exist" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.576672 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.579844 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.586447 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.586664 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.586701 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.586802 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-hh57v" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.586879 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.616150 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.635094 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-internal-tls-certs\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.635159 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-public-tls-certs\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.635208 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-scripts\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.635834 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.635914 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-config-data\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.636114 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfwxc\" (UniqueName: \"kubernetes.io/projected/e2935140-950a-4c06-adee-a72196dec3d8-kube-api-access-mfwxc\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.738325 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-internal-tls-certs\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.738394 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-public-tls-certs\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.738442 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-scripts\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.738576 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.738629 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-config-data\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.738664 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfwxc\" (UniqueName: \"kubernetes.io/projected/e2935140-950a-4c06-adee-a72196dec3d8-kube-api-access-mfwxc\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.742281 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-internal-tls-certs\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.742531 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-config-data\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.744655 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-public-tls-certs\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.744685 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.746167 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2935140-950a-4c06-adee-a72196dec3d8-scripts\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.754847 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfwxc\" (UniqueName: \"kubernetes.io/projected/e2935140-950a-4c06-adee-a72196dec3d8-kube-api-access-mfwxc\") pod \"aodh-0\" (UID: \"e2935140-950a-4c06-adee-a72196dec3d8\") " pod="openstack/aodh-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.815019 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 17:30:33 crc kubenswrapper[4954]: I1209 17:30:33.962462 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 09 17:30:34 crc kubenswrapper[4954]: I1209 17:30:34.138511 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93f122ae-71b4-4393-a163-9f61a9b52b8d" path="/var/lib/kubelet/pods/93f122ae-71b4-4393-a163-9f61a9b52b8d/volumes" Dec 09 17:30:34 crc kubenswrapper[4954]: I1209 17:30:34.142483 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c" path="/var/lib/kubelet/pods/f8c4e70e-9b93-42c6-b2e7-ce97ecb4901c/volumes" Dec 09 17:30:34 crc kubenswrapper[4954]: I1209 17:30:34.318270 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 17:30:34 crc kubenswrapper[4954]: W1209 17:30:34.319237 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf299ddad_e1de_4ea5_a76c_35477739f14e.slice/crio-cac6ec146ef54de420d20fee70b7780dce35ba3f748f79fb778d73941c5033e0 WatchSource:0}: Error finding container cac6ec146ef54de420d20fee70b7780dce35ba3f748f79fb778d73941c5033e0: Status 404 returned error can't find the container with id cac6ec146ef54de420d20fee70b7780dce35ba3f748f79fb778d73941c5033e0 Dec 09 17:30:34 crc kubenswrapper[4954]: I1209 17:30:34.530153 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 09 17:30:34 crc kubenswrapper[4954]: W1209 17:30:34.530857 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2935140_950a_4c06_adee_a72196dec3d8.slice/crio-1792e36ce5a4b59ad4c7431344467d288da438208911a6760b2f0f21e57c5c4b WatchSource:0}: Error finding container 1792e36ce5a4b59ad4c7431344467d288da438208911a6760b2f0f21e57c5c4b: Status 404 returned error can't find the container with id 1792e36ce5a4b59ad4c7431344467d288da438208911a6760b2f0f21e57c5c4b Dec 09 17:30:35 crc kubenswrapper[4954]: I1209 17:30:35.239353 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f299ddad-e1de-4ea5-a76c-35477739f14e","Type":"ContainerStarted","Data":"63914c6bdaa07cfe87caadb731c33b7074e0e6abe7022b21073ce25c645f4c67"} Dec 09 17:30:35 crc kubenswrapper[4954]: I1209 17:30:35.239755 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f299ddad-e1de-4ea5-a76c-35477739f14e","Type":"ContainerStarted","Data":"bb4ed6e00fac4bd56384917819046e2059d06e5f5ebb991c05087cdfbb9ea560"} Dec 09 17:30:35 crc kubenswrapper[4954]: I1209 17:30:35.239786 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f299ddad-e1de-4ea5-a76c-35477739f14e","Type":"ContainerStarted","Data":"cac6ec146ef54de420d20fee70b7780dce35ba3f748f79fb778d73941c5033e0"} Dec 09 17:30:35 crc kubenswrapper[4954]: I1209 17:30:35.241772 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2935140-950a-4c06-adee-a72196dec3d8","Type":"ContainerStarted","Data":"1792e36ce5a4b59ad4c7431344467d288da438208911a6760b2f0f21e57c5c4b"} Dec 09 17:30:35 crc kubenswrapper[4954]: I1209 17:30:35.276882 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.276852363 podStartE2EDuration="2.276852363s" podCreationTimestamp="2025-12-09 17:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:30:35.273211359 +0000 UTC m=+2031.661385189" watchObservedRunningTime="2025-12-09 17:30:35.276852363 +0000 UTC m=+2031.665026183" Dec 09 17:30:36 crc kubenswrapper[4954]: I1209 17:30:36.253398 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2935140-950a-4c06-adee-a72196dec3d8","Type":"ContainerStarted","Data":"0efee4e7e347eef11461461620af5a65b78bbf92372312d39110183012384df8"} Dec 09 17:30:36 crc kubenswrapper[4954]: I1209 17:30:36.253769 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2935140-950a-4c06-adee-a72196dec3d8","Type":"ContainerStarted","Data":"8132118a30f889339c7f025c801220462ed3d26ae0c9f6af20fb2bda38860ef6"} Dec 09 17:30:36 crc kubenswrapper[4954]: I1209 17:30:36.494270 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 17:30:38 crc kubenswrapper[4954]: I1209 17:30:38.277718 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2935140-950a-4c06-adee-a72196dec3d8","Type":"ContainerStarted","Data":"5bf0a445a30fc863029808a2f3489a7faccc9a2d7b47e14ed4aa6278af5200c9"} Dec 09 17:30:38 crc kubenswrapper[4954]: I1209 17:30:38.815954 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 17:30:38 crc kubenswrapper[4954]: I1209 17:30:38.816336 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 17:30:39 crc kubenswrapper[4954]: I1209 17:30:39.292206 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2935140-950a-4c06-adee-a72196dec3d8","Type":"ContainerStarted","Data":"90c10cce646a2be07b51701f1399dfbd1afed6cff78948b232c547509feb3386"} Dec 09 17:30:39 crc kubenswrapper[4954]: I1209 17:30:39.329550 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.378211486 podStartE2EDuration="6.329510233s" podCreationTimestamp="2025-12-09 17:30:33 +0000 UTC" firstStartedPulling="2025-12-09 17:30:34.533935476 +0000 UTC m=+2030.922109296" lastFinishedPulling="2025-12-09 17:30:38.485234223 +0000 UTC m=+2034.873408043" observedRunningTime="2025-12-09 17:30:39.315480144 +0000 UTC m=+2035.703653984" watchObservedRunningTime="2025-12-09 17:30:39.329510233 +0000 UTC m=+2035.717684053" Dec 09 17:30:39 crc kubenswrapper[4954]: I1209 17:30:39.794812 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 17:30:39 crc kubenswrapper[4954]: I1209 17:30:39.795181 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 17:30:40 crc kubenswrapper[4954]: I1209 17:30:40.810762 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="01db8384-81a2-4c9d-b26d-a8d788614768" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.244:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 17:30:40 crc kubenswrapper[4954]: I1209 17:30:40.810801 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="01db8384-81a2-4c9d-b26d-a8d788614768" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.244:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 17:30:41 crc kubenswrapper[4954]: I1209 17:30:41.495130 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 17:30:41 crc kubenswrapper[4954]: I1209 17:30:41.535115 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 17:30:42 crc kubenswrapper[4954]: I1209 17:30:42.360539 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 17:30:43 crc kubenswrapper[4954]: I1209 17:30:43.815284 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 17:30:43 crc kubenswrapper[4954]: I1209 17:30:43.815444 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 17:30:44 crc kubenswrapper[4954]: I1209 17:30:44.829938 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f299ddad-e1de-4ea5-a76c-35477739f14e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.246:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 17:30:44 crc kubenswrapper[4954]: I1209 17:30:44.829983 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f299ddad-e1de-4ea5-a76c-35477739f14e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.246:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 17:30:47 crc kubenswrapper[4954]: I1209 17:30:47.432655 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 17:30:49 crc kubenswrapper[4954]: I1209 17:30:49.806925 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 17:30:49 crc kubenswrapper[4954]: I1209 17:30:49.807912 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 17:30:49 crc kubenswrapper[4954]: I1209 17:30:49.816321 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 17:30:49 crc kubenswrapper[4954]: I1209 17:30:49.819861 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 17:30:50 crc kubenswrapper[4954]: I1209 17:30:50.443894 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 17:30:50 crc kubenswrapper[4954]: I1209 17:30:50.450516 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 17:30:52 crc kubenswrapper[4954]: I1209 17:30:52.644884 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 17:30:52 crc kubenswrapper[4954]: I1209 17:30:52.645716 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="72a212eb-2f11-46aa-88ff-6b59b86c5b6c" containerName="kube-state-metrics" containerID="cri-o://bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58" gracePeriod=30 Dec 09 17:30:52 crc kubenswrapper[4954]: I1209 17:30:52.825411 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 17:30:52 crc kubenswrapper[4954]: I1209 17:30:52.829080 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="ce89f21f-56e2-4107-ae7a-cb38603a2e01" containerName="mysqld-exporter" containerID="cri-o://7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4" gracePeriod=30 Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.299754 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.411127 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.461547 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fnnt\" (UniqueName: \"kubernetes.io/projected/72a212eb-2f11-46aa-88ff-6b59b86c5b6c-kube-api-access-9fnnt\") pod \"72a212eb-2f11-46aa-88ff-6b59b86c5b6c\" (UID: \"72a212eb-2f11-46aa-88ff-6b59b86c5b6c\") " Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.471781 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72a212eb-2f11-46aa-88ff-6b59b86c5b6c-kube-api-access-9fnnt" (OuterVolumeSpecName: "kube-api-access-9fnnt") pod "72a212eb-2f11-46aa-88ff-6b59b86c5b6c" (UID: "72a212eb-2f11-46aa-88ff-6b59b86c5b6c"). InnerVolumeSpecName "kube-api-access-9fnnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.477489 4954 generic.go:334] "Generic (PLEG): container finished" podID="72a212eb-2f11-46aa-88ff-6b59b86c5b6c" containerID="bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58" exitCode=2 Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.477579 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"72a212eb-2f11-46aa-88ff-6b59b86c5b6c","Type":"ContainerDied","Data":"bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58"} Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.477637 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"72a212eb-2f11-46aa-88ff-6b59b86c5b6c","Type":"ContainerDied","Data":"cf857834225d1af96ea55cf4172a4be1ac63e1d856ef593388b4a899d1020081"} Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.477661 4954 scope.go:117] "RemoveContainer" containerID="bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.477886 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.480999 4954 generic.go:334] "Generic (PLEG): container finished" podID="ce89f21f-56e2-4107-ae7a-cb38603a2e01" containerID="7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4" exitCode=2 Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.481054 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"ce89f21f-56e2-4107-ae7a-cb38603a2e01","Type":"ContainerDied","Data":"7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4"} Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.481089 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"ce89f21f-56e2-4107-ae7a-cb38603a2e01","Type":"ContainerDied","Data":"563b1afc7a6a0927b41c3a37f441088d67e2657c116c6f1f174dfcc05155e0c0"} Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.481168 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.521906 4954 scope.go:117] "RemoveContainer" containerID="bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58" Dec 09 17:30:53 crc kubenswrapper[4954]: E1209 17:30:53.522423 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58\": container with ID starting with bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58 not found: ID does not exist" containerID="bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.522482 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58"} err="failed to get container status \"bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58\": rpc error: code = NotFound desc = could not find container \"bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58\": container with ID starting with bf7680dde005d32b1704a7bba886161ebb66e05a0cd681a53ebecaecc4102a58 not found: ID does not exist" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.522516 4954 scope.go:117] "RemoveContainer" containerID="7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.524749 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.554882 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.565916 4954 scope.go:117] "RemoveContainer" containerID="7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4" Dec 09 17:30:53 crc kubenswrapper[4954]: E1209 17:30:53.566786 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4\": container with ID starting with 7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4 not found: ID does not exist" containerID="7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.566858 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4"} err="failed to get container status \"7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4\": rpc error: code = NotFound desc = could not find container \"7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4\": container with ID starting with 7cf9fb67a94fbf7afcd430eb0ec5a3679c79326ce343d7391181f1b0e2e5dfe4 not found: ID does not exist" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.566827 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-combined-ca-bundle\") pod \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.567167 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-config-data\") pod \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.567348 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rwtc\" (UniqueName: \"kubernetes.io/projected/ce89f21f-56e2-4107-ae7a-cb38603a2e01-kube-api-access-8rwtc\") pod \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\" (UID: \"ce89f21f-56e2-4107-ae7a-cb38603a2e01\") " Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.568154 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fnnt\" (UniqueName: \"kubernetes.io/projected/72a212eb-2f11-46aa-88ff-6b59b86c5b6c-kube-api-access-9fnnt\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.579058 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce89f21f-56e2-4107-ae7a-cb38603a2e01-kube-api-access-8rwtc" (OuterVolumeSpecName: "kube-api-access-8rwtc") pod "ce89f21f-56e2-4107-ae7a-cb38603a2e01" (UID: "ce89f21f-56e2-4107-ae7a-cb38603a2e01"). InnerVolumeSpecName "kube-api-access-8rwtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.594557 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 17:30:53 crc kubenswrapper[4954]: E1209 17:30:53.595313 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a212eb-2f11-46aa-88ff-6b59b86c5b6c" containerName="kube-state-metrics" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.595340 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a212eb-2f11-46aa-88ff-6b59b86c5b6c" containerName="kube-state-metrics" Dec 09 17:30:53 crc kubenswrapper[4954]: E1209 17:30:53.595363 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce89f21f-56e2-4107-ae7a-cb38603a2e01" containerName="mysqld-exporter" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.595372 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce89f21f-56e2-4107-ae7a-cb38603a2e01" containerName="mysqld-exporter" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.595675 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="72a212eb-2f11-46aa-88ff-6b59b86c5b6c" containerName="kube-state-metrics" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.595701 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce89f21f-56e2-4107-ae7a-cb38603a2e01" containerName="mysqld-exporter" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.596642 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.600270 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.601300 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.624738 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce89f21f-56e2-4107-ae7a-cb38603a2e01" (UID: "ce89f21f-56e2-4107-ae7a-cb38603a2e01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.629715 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.661761 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-config-data" (OuterVolumeSpecName: "config-data") pod "ce89f21f-56e2-4107-ae7a-cb38603a2e01" (UID: "ce89f21f-56e2-4107-ae7a-cb38603a2e01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.670136 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.670180 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce89f21f-56e2-4107-ae7a-cb38603a2e01-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.670193 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rwtc\" (UniqueName: \"kubernetes.io/projected/ce89f21f-56e2-4107-ae7a-cb38603a2e01-kube-api-access-8rwtc\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.771984 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03899af3-5963-4f63-8c51-8fcbbc8582f0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.772064 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/03899af3-5963-4f63-8c51-8fcbbc8582f0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.772330 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqqrc\" (UniqueName: \"kubernetes.io/projected/03899af3-5963-4f63-8c51-8fcbbc8582f0-kube-api-access-kqqrc\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.772386 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/03899af3-5963-4f63-8c51-8fcbbc8582f0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.822660 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.822768 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.826852 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.827796 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.863206 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.877022 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/03899af3-5963-4f63-8c51-8fcbbc8582f0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.877075 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03899af3-5963-4f63-8c51-8fcbbc8582f0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.877149 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqqrc\" (UniqueName: \"kubernetes.io/projected/03899af3-5963-4f63-8c51-8fcbbc8582f0-kube-api-access-kqqrc\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.877174 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/03899af3-5963-4f63-8c51-8fcbbc8582f0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.882512 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/03899af3-5963-4f63-8c51-8fcbbc8582f0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.883462 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.883530 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03899af3-5963-4f63-8c51-8fcbbc8582f0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.883834 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/03899af3-5963-4f63-8c51-8fcbbc8582f0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.896791 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.898960 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.901632 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.901944 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.919825 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqqrc\" (UniqueName: \"kubernetes.io/projected/03899af3-5963-4f63-8c51-8fcbbc8582f0-kube-api-access-kqqrc\") pod \"kube-state-metrics-0\" (UID: \"03899af3-5963-4f63-8c51-8fcbbc8582f0\") " pod="openstack/kube-state-metrics-0" Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.940363 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 17:30:53 crc kubenswrapper[4954]: I1209 17:30:53.959860 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.127962 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmjnh\" (UniqueName: \"kubernetes.io/projected/9785110a-93dc-4330-8198-792183afdb87-kube-api-access-wmjnh\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.128024 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/9785110a-93dc-4330-8198-792183afdb87-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.128103 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9785110a-93dc-4330-8198-792183afdb87-config-data\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.128138 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9785110a-93dc-4330-8198-792183afdb87-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.220735 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72a212eb-2f11-46aa-88ff-6b59b86c5b6c" path="/var/lib/kubelet/pods/72a212eb-2f11-46aa-88ff-6b59b86c5b6c/volumes" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.223039 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce89f21f-56e2-4107-ae7a-cb38603a2e01" path="/var/lib/kubelet/pods/ce89f21f-56e2-4107-ae7a-cb38603a2e01/volumes" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.232836 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9785110a-93dc-4330-8198-792183afdb87-config-data\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.232924 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9785110a-93dc-4330-8198-792183afdb87-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.233090 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmjnh\" (UniqueName: \"kubernetes.io/projected/9785110a-93dc-4330-8198-792183afdb87-kube-api-access-wmjnh\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.233138 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/9785110a-93dc-4330-8198-792183afdb87-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.272752 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9785110a-93dc-4330-8198-792183afdb87-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.278248 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9785110a-93dc-4330-8198-792183afdb87-config-data\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.286203 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/9785110a-93dc-4330-8198-792183afdb87-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.293445 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmjnh\" (UniqueName: \"kubernetes.io/projected/9785110a-93dc-4330-8198-792183afdb87-kube-api-access-wmjnh\") pod \"mysqld-exporter-0\" (UID: \"9785110a-93dc-4330-8198-792183afdb87\") " pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.521179 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.769236 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 17:30:54 crc kubenswrapper[4954]: I1209 17:30:54.777793 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 17:30:55 crc kubenswrapper[4954]: I1209 17:30:55.053005 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 09 17:30:55 crc kubenswrapper[4954]: I1209 17:30:55.515954 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"9785110a-93dc-4330-8198-792183afdb87","Type":"ContainerStarted","Data":"2c6727775bc8cc9260294ad3b2662027b160d514ac90bc28870e513539031cf6"} Dec 09 17:30:55 crc kubenswrapper[4954]: I1209 17:30:55.517257 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"03899af3-5963-4f63-8c51-8fcbbc8582f0","Type":"ContainerStarted","Data":"89df3a09ed209373728efd284ac8d03791b9712ee7603ca013b0c1606eceb2b4"} Dec 09 17:30:55 crc kubenswrapper[4954]: I1209 17:30:55.767959 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:55 crc kubenswrapper[4954]: I1209 17:30:55.769304 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="ceilometer-central-agent" containerID="cri-o://d237c59ecd021445baf479358507e0aef7babd5f0631c0c91db2eda378dfbbdf" gracePeriod=30 Dec 09 17:30:55 crc kubenswrapper[4954]: I1209 17:30:55.769837 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="ceilometer-notification-agent" containerID="cri-o://ff078a7a54cf55d551ed7b1cc84fd4dc3a7ee136d9797be4beb31a5c0b4022f6" gracePeriod=30 Dec 09 17:30:55 crc kubenswrapper[4954]: I1209 17:30:55.769821 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="proxy-httpd" containerID="cri-o://d9e65b648349a288fb93f22404ff0b315eeca9b04e2cd46e7428d85e95c4132d" gracePeriod=30 Dec 09 17:30:55 crc kubenswrapper[4954]: I1209 17:30:55.769937 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="sg-core" containerID="cri-o://4be6620af5968a980115be19bd7c8ebe122982fe16e90eb212a05e310652bf55" gracePeriod=30 Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.540251 4954 generic.go:334] "Generic (PLEG): container finished" podID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerID="d9e65b648349a288fb93f22404ff0b315eeca9b04e2cd46e7428d85e95c4132d" exitCode=0 Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.540616 4954 generic.go:334] "Generic (PLEG): container finished" podID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerID="4be6620af5968a980115be19bd7c8ebe122982fe16e90eb212a05e310652bf55" exitCode=2 Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.540630 4954 generic.go:334] "Generic (PLEG): container finished" podID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerID="d237c59ecd021445baf479358507e0aef7babd5f0631c0c91db2eda378dfbbdf" exitCode=0 Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.540771 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1df5925-6ceb-466c-a3bd-0b3373c7351d","Type":"ContainerDied","Data":"d9e65b648349a288fb93f22404ff0b315eeca9b04e2cd46e7428d85e95c4132d"} Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.540811 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1df5925-6ceb-466c-a3bd-0b3373c7351d","Type":"ContainerDied","Data":"4be6620af5968a980115be19bd7c8ebe122982fe16e90eb212a05e310652bf55"} Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.540846 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1df5925-6ceb-466c-a3bd-0b3373c7351d","Type":"ContainerDied","Data":"d237c59ecd021445baf479358507e0aef7babd5f0631c0c91db2eda378dfbbdf"} Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.544605 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"9785110a-93dc-4330-8198-792183afdb87","Type":"ContainerStarted","Data":"259f2f476812a2e63514ef624bc47dbd18356b7c3d9e2e05a75af1f622823492"} Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.551002 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"03899af3-5963-4f63-8c51-8fcbbc8582f0","Type":"ContainerStarted","Data":"50a1c98744084683ebff7abe9d20eccece649d45a48dc5a0084bca5c86cb82ea"} Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.551163 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.564752 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.949977644 podStartE2EDuration="3.564711089s" podCreationTimestamp="2025-12-09 17:30:53 +0000 UTC" firstStartedPulling="2025-12-09 17:30:55.06533455 +0000 UTC m=+2051.453508370" lastFinishedPulling="2025-12-09 17:30:55.680067995 +0000 UTC m=+2052.068241815" observedRunningTime="2025-12-09 17:30:56.563099798 +0000 UTC m=+2052.951273618" watchObservedRunningTime="2025-12-09 17:30:56.564711089 +0000 UTC m=+2052.952884909" Dec 09 17:30:56 crc kubenswrapper[4954]: I1209 17:30:56.593902 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.970231117 podStartE2EDuration="3.593872161s" podCreationTimestamp="2025-12-09 17:30:53 +0000 UTC" firstStartedPulling="2025-12-09 17:30:54.769021284 +0000 UTC m=+2051.157195104" lastFinishedPulling="2025-12-09 17:30:55.392662328 +0000 UTC m=+2051.780836148" observedRunningTime="2025-12-09 17:30:56.586335866 +0000 UTC m=+2052.974509696" watchObservedRunningTime="2025-12-09 17:30:56.593872161 +0000 UTC m=+2052.982045981" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.569611 4954 generic.go:334] "Generic (PLEG): container finished" podID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerID="ff078a7a54cf55d551ed7b1cc84fd4dc3a7ee136d9797be4beb31a5c0b4022f6" exitCode=0 Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.569818 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1df5925-6ceb-466c-a3bd-0b3373c7351d","Type":"ContainerDied","Data":"ff078a7a54cf55d551ed7b1cc84fd4dc3a7ee136d9797be4beb31a5c0b4022f6"} Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.571074 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1df5925-6ceb-466c-a3bd-0b3373c7351d","Type":"ContainerDied","Data":"4120eb48ba18b7df26a4e7187992d7e0665c6cd463453be0036b3604ebc6093f"} Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.571090 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4120eb48ba18b7df26a4e7187992d7e0665c6cd463453be0036b3604ebc6093f" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.648412 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.814257 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-combined-ca-bundle\") pod \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.814717 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwslc\" (UniqueName: \"kubernetes.io/projected/a1df5925-6ceb-466c-a3bd-0b3373c7351d-kube-api-access-zwslc\") pod \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.814830 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-sg-core-conf-yaml\") pod \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.814968 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-config-data\") pod \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.815061 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-scripts\") pod \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.815095 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-log-httpd\") pod \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.815122 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-run-httpd\") pod \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\" (UID: \"a1df5925-6ceb-466c-a3bd-0b3373c7351d\") " Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.815699 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a1df5925-6ceb-466c-a3bd-0b3373c7351d" (UID: "a1df5925-6ceb-466c-a3bd-0b3373c7351d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.815819 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a1df5925-6ceb-466c-a3bd-0b3373c7351d" (UID: "a1df5925-6ceb-466c-a3bd-0b3373c7351d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.816658 4954 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.816684 4954 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1df5925-6ceb-466c-a3bd-0b3373c7351d-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.820547 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-scripts" (OuterVolumeSpecName: "scripts") pod "a1df5925-6ceb-466c-a3bd-0b3373c7351d" (UID: "a1df5925-6ceb-466c-a3bd-0b3373c7351d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.821741 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1df5925-6ceb-466c-a3bd-0b3373c7351d-kube-api-access-zwslc" (OuterVolumeSpecName: "kube-api-access-zwslc") pod "a1df5925-6ceb-466c-a3bd-0b3373c7351d" (UID: "a1df5925-6ceb-466c-a3bd-0b3373c7351d"). InnerVolumeSpecName "kube-api-access-zwslc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.859768 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a1df5925-6ceb-466c-a3bd-0b3373c7351d" (UID: "a1df5925-6ceb-466c-a3bd-0b3373c7351d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.907454 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1df5925-6ceb-466c-a3bd-0b3373c7351d" (UID: "a1df5925-6ceb-466c-a3bd-0b3373c7351d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.919630 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.919677 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.919692 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwslc\" (UniqueName: \"kubernetes.io/projected/a1df5925-6ceb-466c-a3bd-0b3373c7351d-kube-api-access-zwslc\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.919704 4954 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:57 crc kubenswrapper[4954]: I1209 17:30:57.944406 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-config-data" (OuterVolumeSpecName: "config-data") pod "a1df5925-6ceb-466c-a3bd-0b3373c7351d" (UID: "a1df5925-6ceb-466c-a3bd-0b3373c7351d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.021648 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1df5925-6ceb-466c-a3bd-0b3373c7351d-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.581407 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.659139 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.677945 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.691465 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:58 crc kubenswrapper[4954]: E1209 17:30:58.692255 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="ceilometer-notification-agent" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.692325 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="ceilometer-notification-agent" Dec 09 17:30:58 crc kubenswrapper[4954]: E1209 17:30:58.692348 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="ceilometer-central-agent" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.692357 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="ceilometer-central-agent" Dec 09 17:30:58 crc kubenswrapper[4954]: E1209 17:30:58.692375 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="sg-core" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.692384 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="sg-core" Dec 09 17:30:58 crc kubenswrapper[4954]: E1209 17:30:58.692398 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="proxy-httpd" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.692406 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="proxy-httpd" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.692764 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="sg-core" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.692788 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="ceilometer-notification-agent" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.692809 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="proxy-httpd" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.692824 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" containerName="ceilometer-central-agent" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.695774 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.699562 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.699973 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.700838 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.712506 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.837856 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-run-httpd\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.837959 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.838010 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.838042 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-log-httpd\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.838076 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5ckv\" (UniqueName: \"kubernetes.io/projected/940c8621-fab6-4945-a800-aec88957e8cd-kube-api-access-j5ckv\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.838118 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-scripts\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.838159 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.838196 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-config-data\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.940439 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.940811 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.940852 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-log-httpd\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.940893 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5ckv\" (UniqueName: \"kubernetes.io/projected/940c8621-fab6-4945-a800-aec88957e8cd-kube-api-access-j5ckv\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.940935 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-scripts\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.940971 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.941008 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-config-data\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.941050 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-run-httpd\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.941525 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-run-httpd\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.942343 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-log-httpd\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.945165 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.945301 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-scripts\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.946549 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-config-data\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.948144 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.950544 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:58 crc kubenswrapper[4954]: I1209 17:30:58.964476 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5ckv\" (UniqueName: \"kubernetes.io/projected/940c8621-fab6-4945-a800-aec88957e8cd-kube-api-access-j5ckv\") pod \"ceilometer-0\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " pod="openstack/ceilometer-0" Dec 09 17:30:59 crc kubenswrapper[4954]: I1209 17:30:59.015213 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:30:59 crc kubenswrapper[4954]: I1209 17:30:59.568832 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:30:59 crc kubenswrapper[4954]: I1209 17:30:59.602489 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"940c8621-fab6-4945-a800-aec88957e8cd","Type":"ContainerStarted","Data":"84ae3c56cb4e17e74eff60794d20cd1aa7926528fe6cc8b8e34126c8fbc2808f"} Dec 09 17:31:00 crc kubenswrapper[4954]: I1209 17:31:00.135071 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1df5925-6ceb-466c-a3bd-0b3373c7351d" path="/var/lib/kubelet/pods/a1df5925-6ceb-466c-a3bd-0b3373c7351d/volumes" Dec 09 17:31:01 crc kubenswrapper[4954]: I1209 17:31:01.625396 4954 scope.go:117] "RemoveContainer" containerID="3ae33c49509f232f53e0192aa376275f6eeeac6124ea372e9313da830898fc77" Dec 09 17:31:01 crc kubenswrapper[4954]: I1209 17:31:01.755316 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vkrb7"] Dec 09 17:31:01 crc kubenswrapper[4954]: I1209 17:31:01.762874 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:01 crc kubenswrapper[4954]: I1209 17:31:01.776449 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vkrb7"] Dec 09 17:31:01 crc kubenswrapper[4954]: I1209 17:31:01.931805 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nr66\" (UniqueName: \"kubernetes.io/projected/50f2cac6-07a1-460d-a2c6-520120d32d7a-kube-api-access-4nr66\") pod \"redhat-operators-vkrb7\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:01 crc kubenswrapper[4954]: I1209 17:31:01.931961 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-catalog-content\") pod \"redhat-operators-vkrb7\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:01 crc kubenswrapper[4954]: I1209 17:31:01.932077 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-utilities\") pod \"redhat-operators-vkrb7\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.034249 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-catalog-content\") pod \"redhat-operators-vkrb7\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.034475 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-utilities\") pod \"redhat-operators-vkrb7\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.034749 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nr66\" (UniqueName: \"kubernetes.io/projected/50f2cac6-07a1-460d-a2c6-520120d32d7a-kube-api-access-4nr66\") pod \"redhat-operators-vkrb7\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.036053 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-catalog-content\") pod \"redhat-operators-vkrb7\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.036350 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-utilities\") pod \"redhat-operators-vkrb7\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.054538 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nr66\" (UniqueName: \"kubernetes.io/projected/50f2cac6-07a1-460d-a2c6-520120d32d7a-kube-api-access-4nr66\") pod \"redhat-operators-vkrb7\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.098231 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.602790 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vkrb7"] Dec 09 17:31:02 crc kubenswrapper[4954]: W1209 17:31:02.615053 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50f2cac6_07a1_460d_a2c6_520120d32d7a.slice/crio-dbef3726712af7b3a9cdd0c4618e08e0da5153bdbb514b87d3cd816d42aa3c09 WatchSource:0}: Error finding container dbef3726712af7b3a9cdd0c4618e08e0da5153bdbb514b87d3cd816d42aa3c09: Status 404 returned error can't find the container with id dbef3726712af7b3a9cdd0c4618e08e0da5153bdbb514b87d3cd816d42aa3c09 Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.639327 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"940c8621-fab6-4945-a800-aec88957e8cd","Type":"ContainerStarted","Data":"da7b83d3a6a545bee71437eefbc7b8ca7de1f4612a721d1bcd5cfefabe2f962d"} Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.639387 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"940c8621-fab6-4945-a800-aec88957e8cd","Type":"ContainerStarted","Data":"8f36510b924347a56363939c1d38bac1eea2f327fd91b2bd13ec80f76744fa5e"} Dec 09 17:31:02 crc kubenswrapper[4954]: I1209 17:31:02.642629 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrb7" event={"ID":"50f2cac6-07a1-460d-a2c6-520120d32d7a","Type":"ContainerStarted","Data":"dbef3726712af7b3a9cdd0c4618e08e0da5153bdbb514b87d3cd816d42aa3c09"} Dec 09 17:31:03 crc kubenswrapper[4954]: I1209 17:31:03.657892 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"940c8621-fab6-4945-a800-aec88957e8cd","Type":"ContainerStarted","Data":"966968071c3dca458b02e9bddf7d144af549e78eaa2eda6d758bc22aac2d1b3b"} Dec 09 17:31:03 crc kubenswrapper[4954]: I1209 17:31:03.661640 4954 generic.go:334] "Generic (PLEG): container finished" podID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerID="c879f5498c3c27cc1f3d427d7365d198f7ee49a2940726bd4e52bcb75cdae795" exitCode=0 Dec 09 17:31:03 crc kubenswrapper[4954]: I1209 17:31:03.661714 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrb7" event={"ID":"50f2cac6-07a1-460d-a2c6-520120d32d7a","Type":"ContainerDied","Data":"c879f5498c3c27cc1f3d427d7365d198f7ee49a2940726bd4e52bcb75cdae795"} Dec 09 17:31:03 crc kubenswrapper[4954]: I1209 17:31:03.974609 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 09 17:31:04 crc kubenswrapper[4954]: I1209 17:31:04.675416 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrb7" event={"ID":"50f2cac6-07a1-460d-a2c6-520120d32d7a","Type":"ContainerStarted","Data":"a31e6830c8c4e85be6d104a12e0cb31faed31356bb5b95e1efe92f5d618b46e4"} Dec 09 17:31:05 crc kubenswrapper[4954]: I1209 17:31:05.692928 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"940c8621-fab6-4945-a800-aec88957e8cd","Type":"ContainerStarted","Data":"2c73861d780e5a6274dd7b05f6c3cf01b8f5a1544088efbd02325052ce510258"} Dec 09 17:31:05 crc kubenswrapper[4954]: I1209 17:31:05.693264 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 17:31:05 crc kubenswrapper[4954]: I1209 17:31:05.727029 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.650488537 podStartE2EDuration="7.726998959s" podCreationTimestamp="2025-12-09 17:30:58 +0000 UTC" firstStartedPulling="2025-12-09 17:30:59.56963766 +0000 UTC m=+2055.957811480" lastFinishedPulling="2025-12-09 17:31:04.646148082 +0000 UTC m=+2061.034321902" observedRunningTime="2025-12-09 17:31:05.721055563 +0000 UTC m=+2062.109229393" watchObservedRunningTime="2025-12-09 17:31:05.726998959 +0000 UTC m=+2062.115172809" Dec 09 17:31:08 crc kubenswrapper[4954]: I1209 17:31:08.721707 4954 generic.go:334] "Generic (PLEG): container finished" podID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerID="a31e6830c8c4e85be6d104a12e0cb31faed31356bb5b95e1efe92f5d618b46e4" exitCode=0 Dec 09 17:31:08 crc kubenswrapper[4954]: I1209 17:31:08.721761 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrb7" event={"ID":"50f2cac6-07a1-460d-a2c6-520120d32d7a","Type":"ContainerDied","Data":"a31e6830c8c4e85be6d104a12e0cb31faed31356bb5b95e1efe92f5d618b46e4"} Dec 09 17:31:09 crc kubenswrapper[4954]: I1209 17:31:09.736255 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrb7" event={"ID":"50f2cac6-07a1-460d-a2c6-520120d32d7a","Type":"ContainerStarted","Data":"c755572a654f5173f3a95e07780da6944242cadbd21de04cf1cb451d31f62cc9"} Dec 09 17:31:09 crc kubenswrapper[4954]: I1209 17:31:09.764952 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vkrb7" podStartSLOduration=3.249965419 podStartE2EDuration="8.764922778s" podCreationTimestamp="2025-12-09 17:31:01 +0000 UTC" firstStartedPulling="2025-12-09 17:31:03.664838372 +0000 UTC m=+2060.053012192" lastFinishedPulling="2025-12-09 17:31:09.179795731 +0000 UTC m=+2065.567969551" observedRunningTime="2025-12-09 17:31:09.755429811 +0000 UTC m=+2066.143603651" watchObservedRunningTime="2025-12-09 17:31:09.764922778 +0000 UTC m=+2066.153096598" Dec 09 17:31:12 crc kubenswrapper[4954]: I1209 17:31:12.098936 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:12 crc kubenswrapper[4954]: I1209 17:31:12.099464 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:13 crc kubenswrapper[4954]: I1209 17:31:13.168665 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vkrb7" podUID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerName="registry-server" probeResult="failure" output=< Dec 09 17:31:13 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 17:31:13 crc kubenswrapper[4954]: > Dec 09 17:31:22 crc kubenswrapper[4954]: I1209 17:31:22.149467 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:22 crc kubenswrapper[4954]: I1209 17:31:22.200615 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:24 crc kubenswrapper[4954]: I1209 17:31:24.729982 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vkrb7"] Dec 09 17:31:24 crc kubenswrapper[4954]: I1209 17:31:24.730538 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vkrb7" podUID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerName="registry-server" containerID="cri-o://c755572a654f5173f3a95e07780da6944242cadbd21de04cf1cb451d31f62cc9" gracePeriod=2 Dec 09 17:31:24 crc kubenswrapper[4954]: I1209 17:31:24.927454 4954 generic.go:334] "Generic (PLEG): container finished" podID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerID="c755572a654f5173f3a95e07780da6944242cadbd21de04cf1cb451d31f62cc9" exitCode=0 Dec 09 17:31:24 crc kubenswrapper[4954]: I1209 17:31:24.927517 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrb7" event={"ID":"50f2cac6-07a1-460d-a2c6-520120d32d7a","Type":"ContainerDied","Data":"c755572a654f5173f3a95e07780da6944242cadbd21de04cf1cb451d31f62cc9"} Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.418060 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.496781 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-utilities\") pod \"50f2cac6-07a1-460d-a2c6-520120d32d7a\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.496902 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-catalog-content\") pod \"50f2cac6-07a1-460d-a2c6-520120d32d7a\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.497047 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nr66\" (UniqueName: \"kubernetes.io/projected/50f2cac6-07a1-460d-a2c6-520120d32d7a-kube-api-access-4nr66\") pod \"50f2cac6-07a1-460d-a2c6-520120d32d7a\" (UID: \"50f2cac6-07a1-460d-a2c6-520120d32d7a\") " Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.497764 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-utilities" (OuterVolumeSpecName: "utilities") pod "50f2cac6-07a1-460d-a2c6-520120d32d7a" (UID: "50f2cac6-07a1-460d-a2c6-520120d32d7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.504671 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50f2cac6-07a1-460d-a2c6-520120d32d7a-kube-api-access-4nr66" (OuterVolumeSpecName: "kube-api-access-4nr66") pod "50f2cac6-07a1-460d-a2c6-520120d32d7a" (UID: "50f2cac6-07a1-460d-a2c6-520120d32d7a"). InnerVolumeSpecName "kube-api-access-4nr66". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.599639 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nr66\" (UniqueName: \"kubernetes.io/projected/50f2cac6-07a1-460d-a2c6-520120d32d7a-kube-api-access-4nr66\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.600048 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.625308 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50f2cac6-07a1-460d-a2c6-520120d32d7a" (UID: "50f2cac6-07a1-460d-a2c6-520120d32d7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.702397 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50f2cac6-07a1-460d-a2c6-520120d32d7a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.942247 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkrb7" event={"ID":"50f2cac6-07a1-460d-a2c6-520120d32d7a","Type":"ContainerDied","Data":"dbef3726712af7b3a9cdd0c4618e08e0da5153bdbb514b87d3cd816d42aa3c09"} Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.942320 4954 scope.go:117] "RemoveContainer" containerID="c755572a654f5173f3a95e07780da6944242cadbd21de04cf1cb451d31f62cc9" Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.942417 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkrb7" Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.982839 4954 scope.go:117] "RemoveContainer" containerID="a31e6830c8c4e85be6d104a12e0cb31faed31356bb5b95e1efe92f5d618b46e4" Dec 09 17:31:25 crc kubenswrapper[4954]: I1209 17:31:25.990821 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vkrb7"] Dec 09 17:31:26 crc kubenswrapper[4954]: I1209 17:31:26.003720 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vkrb7"] Dec 09 17:31:26 crc kubenswrapper[4954]: I1209 17:31:26.014319 4954 scope.go:117] "RemoveContainer" containerID="c879f5498c3c27cc1f3d427d7365d198f7ee49a2940726bd4e52bcb75cdae795" Dec 09 17:31:26 crc kubenswrapper[4954]: I1209 17:31:26.135294 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50f2cac6-07a1-460d-a2c6-520120d32d7a" path="/var/lib/kubelet/pods/50f2cac6-07a1-460d-a2c6-520120d32d7a/volumes" Dec 09 17:31:29 crc kubenswrapper[4954]: I1209 17:31:29.027121 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.521390 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-p5ndh"] Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.535221 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-p5ndh"] Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.661774 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-b2pg9"] Dec 09 17:31:41 crc kubenswrapper[4954]: E1209 17:31:41.662355 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerName="extract-content" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.662377 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerName="extract-content" Dec 09 17:31:41 crc kubenswrapper[4954]: E1209 17:31:41.662393 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerName="registry-server" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.662398 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerName="registry-server" Dec 09 17:31:41 crc kubenswrapper[4954]: E1209 17:31:41.662423 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerName="extract-utilities" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.662430 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerName="extract-utilities" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.662683 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="50f2cac6-07a1-460d-a2c6-520120d32d7a" containerName="registry-server" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.663828 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.684739 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-b2pg9"] Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.696976 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz7bq\" (UniqueName: \"kubernetes.io/projected/8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38-kube-api-access-rz7bq\") pod \"heat-db-sync-b2pg9\" (UID: \"8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38\") " pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.697050 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38-combined-ca-bundle\") pod \"heat-db-sync-b2pg9\" (UID: \"8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38\") " pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.697253 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38-config-data\") pod \"heat-db-sync-b2pg9\" (UID: \"8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38\") " pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.800489 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz7bq\" (UniqueName: \"kubernetes.io/projected/8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38-kube-api-access-rz7bq\") pod \"heat-db-sync-b2pg9\" (UID: \"8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38\") " pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.800561 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38-combined-ca-bundle\") pod \"heat-db-sync-b2pg9\" (UID: \"8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38\") " pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.800663 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38-config-data\") pod \"heat-db-sync-b2pg9\" (UID: \"8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38\") " pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.807730 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38-config-data\") pod \"heat-db-sync-b2pg9\" (UID: \"8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38\") " pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.808397 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38-combined-ca-bundle\") pod \"heat-db-sync-b2pg9\" (UID: \"8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38\") " pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.821126 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz7bq\" (UniqueName: \"kubernetes.io/projected/8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38-kube-api-access-rz7bq\") pod \"heat-db-sync-b2pg9\" (UID: \"8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38\") " pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:41 crc kubenswrapper[4954]: I1209 17:31:41.989451 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-b2pg9" Dec 09 17:31:42 crc kubenswrapper[4954]: I1209 17:31:42.154919 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25667d18-6643-4344-9ac3-e3a4f1af6957" path="/var/lib/kubelet/pods/25667d18-6643-4344-9ac3-e3a4f1af6957/volumes" Dec 09 17:31:42 crc kubenswrapper[4954]: I1209 17:31:42.652945 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-b2pg9"] Dec 09 17:31:42 crc kubenswrapper[4954]: E1209 17:31:42.810675 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:31:42 crc kubenswrapper[4954]: E1209 17:31:42.810755 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:31:42 crc kubenswrapper[4954]: E1209 17:31:42.810961 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:31:42 crc kubenswrapper[4954]: E1209 17:31:42.812282 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:31:43 crc kubenswrapper[4954]: I1209 17:31:43.143582 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-b2pg9" event={"ID":"8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38","Type":"ContainerStarted","Data":"29bd7dff95814d93a99d3df1edd11ebdbd6516464979b2fa945b7f4ab81902d1"} Dec 09 17:31:43 crc kubenswrapper[4954]: E1209 17:31:43.146166 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:31:43 crc kubenswrapper[4954]: I1209 17:31:43.791401 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 17:31:44 crc kubenswrapper[4954]: E1209 17:31:44.192133 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:31:44 crc kubenswrapper[4954]: I1209 17:31:44.291642 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:31:44 crc kubenswrapper[4954]: I1209 17:31:44.292433 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="ceilometer-central-agent" containerID="cri-o://8f36510b924347a56363939c1d38bac1eea2f327fd91b2bd13ec80f76744fa5e" gracePeriod=30 Dec 09 17:31:44 crc kubenswrapper[4954]: I1209 17:31:44.293724 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="proxy-httpd" containerID="cri-o://2c73861d780e5a6274dd7b05f6c3cf01b8f5a1544088efbd02325052ce510258" gracePeriod=30 Dec 09 17:31:44 crc kubenswrapper[4954]: I1209 17:31:44.293905 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="ceilometer-notification-agent" containerID="cri-o://da7b83d3a6a545bee71437eefbc7b8ca7de1f4612a721d1bcd5cfefabe2f962d" gracePeriod=30 Dec 09 17:31:44 crc kubenswrapper[4954]: I1209 17:31:44.293960 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="sg-core" containerID="cri-o://966968071c3dca458b02e9bddf7d144af549e78eaa2eda6d758bc22aac2d1b3b" gracePeriod=30 Dec 09 17:31:44 crc kubenswrapper[4954]: I1209 17:31:44.794317 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 17:31:45 crc kubenswrapper[4954]: I1209 17:31:45.178794 4954 generic.go:334] "Generic (PLEG): container finished" podID="940c8621-fab6-4945-a800-aec88957e8cd" containerID="2c73861d780e5a6274dd7b05f6c3cf01b8f5a1544088efbd02325052ce510258" exitCode=0 Dec 09 17:31:45 crc kubenswrapper[4954]: I1209 17:31:45.179204 4954 generic.go:334] "Generic (PLEG): container finished" podID="940c8621-fab6-4945-a800-aec88957e8cd" containerID="966968071c3dca458b02e9bddf7d144af549e78eaa2eda6d758bc22aac2d1b3b" exitCode=2 Dec 09 17:31:45 crc kubenswrapper[4954]: I1209 17:31:45.179290 4954 generic.go:334] "Generic (PLEG): container finished" podID="940c8621-fab6-4945-a800-aec88957e8cd" containerID="8f36510b924347a56363939c1d38bac1eea2f327fd91b2bd13ec80f76744fa5e" exitCode=0 Dec 09 17:31:45 crc kubenswrapper[4954]: I1209 17:31:45.179373 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"940c8621-fab6-4945-a800-aec88957e8cd","Type":"ContainerDied","Data":"2c73861d780e5a6274dd7b05f6c3cf01b8f5a1544088efbd02325052ce510258"} Dec 09 17:31:45 crc kubenswrapper[4954]: I1209 17:31:45.179458 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"940c8621-fab6-4945-a800-aec88957e8cd","Type":"ContainerDied","Data":"966968071c3dca458b02e9bddf7d144af549e78eaa2eda6d758bc22aac2d1b3b"} Dec 09 17:31:45 crc kubenswrapper[4954]: I1209 17:31:45.179523 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"940c8621-fab6-4945-a800-aec88957e8cd","Type":"ContainerDied","Data":"8f36510b924347a56363939c1d38bac1eea2f327fd91b2bd13ec80f76744fa5e"} Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.218977 4954 generic.go:334] "Generic (PLEG): container finished" podID="940c8621-fab6-4945-a800-aec88957e8cd" containerID="da7b83d3a6a545bee71437eefbc7b8ca7de1f4612a721d1bcd5cfefabe2f962d" exitCode=0 Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.219072 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"940c8621-fab6-4945-a800-aec88957e8cd","Type":"ContainerDied","Data":"da7b83d3a6a545bee71437eefbc7b8ca7de1f4612a721d1bcd5cfefabe2f962d"} Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.467481 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.614753 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-combined-ca-bundle\") pod \"940c8621-fab6-4945-a800-aec88957e8cd\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.615256 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-sg-core-conf-yaml\") pod \"940c8621-fab6-4945-a800-aec88957e8cd\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.615317 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-config-data\") pod \"940c8621-fab6-4945-a800-aec88957e8cd\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.615526 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5ckv\" (UniqueName: \"kubernetes.io/projected/940c8621-fab6-4945-a800-aec88957e8cd-kube-api-access-j5ckv\") pod \"940c8621-fab6-4945-a800-aec88957e8cd\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.615609 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-log-httpd\") pod \"940c8621-fab6-4945-a800-aec88957e8cd\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.615665 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-ceilometer-tls-certs\") pod \"940c8621-fab6-4945-a800-aec88957e8cd\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.615710 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-scripts\") pod \"940c8621-fab6-4945-a800-aec88957e8cd\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.615755 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-run-httpd\") pod \"940c8621-fab6-4945-a800-aec88957e8cd\" (UID: \"940c8621-fab6-4945-a800-aec88957e8cd\") " Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.616579 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "940c8621-fab6-4945-a800-aec88957e8cd" (UID: "940c8621-fab6-4945-a800-aec88957e8cd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.616850 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "940c8621-fab6-4945-a800-aec88957e8cd" (UID: "940c8621-fab6-4945-a800-aec88957e8cd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.623351 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-scripts" (OuterVolumeSpecName: "scripts") pod "940c8621-fab6-4945-a800-aec88957e8cd" (UID: "940c8621-fab6-4945-a800-aec88957e8cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.625074 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/940c8621-fab6-4945-a800-aec88957e8cd-kube-api-access-j5ckv" (OuterVolumeSpecName: "kube-api-access-j5ckv") pod "940c8621-fab6-4945-a800-aec88957e8cd" (UID: "940c8621-fab6-4945-a800-aec88957e8cd"). InnerVolumeSpecName "kube-api-access-j5ckv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.684386 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "940c8621-fab6-4945-a800-aec88957e8cd" (UID: "940c8621-fab6-4945-a800-aec88957e8cd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.719112 4954 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.719164 4954 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.719183 4954 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.719199 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5ckv\" (UniqueName: \"kubernetes.io/projected/940c8621-fab6-4945-a800-aec88957e8cd-kube-api-access-j5ckv\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.719207 4954 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/940c8621-fab6-4945-a800-aec88957e8cd-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.739159 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "940c8621-fab6-4945-a800-aec88957e8cd" (UID: "940c8621-fab6-4945-a800-aec88957e8cd"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.775076 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "940c8621-fab6-4945-a800-aec88957e8cd" (UID: "940c8621-fab6-4945-a800-aec88957e8cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.796848 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-config-data" (OuterVolumeSpecName: "config-data") pod "940c8621-fab6-4945-a800-aec88957e8cd" (UID: "940c8621-fab6-4945-a800-aec88957e8cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.821901 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.821934 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:47 crc kubenswrapper[4954]: I1209 17:31:47.821946 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/940c8621-fab6-4945-a800-aec88957e8cd-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.270858 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"940c8621-fab6-4945-a800-aec88957e8cd","Type":"ContainerDied","Data":"84ae3c56cb4e17e74eff60794d20cd1aa7926528fe6cc8b8e34126c8fbc2808f"} Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.271324 4954 scope.go:117] "RemoveContainer" containerID="2c73861d780e5a6274dd7b05f6c3cf01b8f5a1544088efbd02325052ce510258" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.271629 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.327689 4954 scope.go:117] "RemoveContainer" containerID="966968071c3dca458b02e9bddf7d144af549e78eaa2eda6d758bc22aac2d1b3b" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.364435 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.384288 4954 scope.go:117] "RemoveContainer" containerID="da7b83d3a6a545bee71437eefbc7b8ca7de1f4612a721d1bcd5cfefabe2f962d" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.391310 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.434939 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:31:48 crc kubenswrapper[4954]: E1209 17:31:48.435586 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="proxy-httpd" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.435627 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="proxy-httpd" Dec 09 17:31:48 crc kubenswrapper[4954]: E1209 17:31:48.435656 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="ceilometer-notification-agent" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.435664 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="ceilometer-notification-agent" Dec 09 17:31:48 crc kubenswrapper[4954]: E1209 17:31:48.435712 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="ceilometer-central-agent" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.435719 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="ceilometer-central-agent" Dec 09 17:31:48 crc kubenswrapper[4954]: E1209 17:31:48.435737 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="sg-core" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.435744 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="sg-core" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.435992 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="ceilometer-notification-agent" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.436030 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="ceilometer-central-agent" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.436055 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="sg-core" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.436069 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="940c8621-fab6-4945-a800-aec88957e8cd" containerName="proxy-httpd" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.439061 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.444286 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.448980 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.450106 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.457238 4954 scope.go:117] "RemoveContainer" containerID="8f36510b924347a56363939c1d38bac1eea2f327fd91b2bd13ec80f76744fa5e" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.563500 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.571184 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-log-httpd\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.571302 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.571442 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.571472 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-run-httpd\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.571515 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlv6x\" (UniqueName: \"kubernetes.io/projected/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-kube-api-access-dlv6x\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.571539 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-scripts\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.571567 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-config-data\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.571586 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.674163 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-log-httpd\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.674267 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.674401 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.674438 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-run-httpd\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.674495 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlv6x\" (UniqueName: \"kubernetes.io/projected/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-kube-api-access-dlv6x\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.674530 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-scripts\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.674566 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-config-data\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.674681 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.674915 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-log-httpd\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.675907 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-run-httpd\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.686708 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.686982 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-scripts\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.688156 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.688886 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-config-data\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.688962 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.695489 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlv6x\" (UniqueName: \"kubernetes.io/projected/9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df-kube-api-access-dlv6x\") pod \"ceilometer-0\" (UID: \"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df\") " pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.765301 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 17:31:48 crc kubenswrapper[4954]: I1209 17:31:48.909489 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="3645ed14-3d97-496e-a683-2e56ea33fec9" containerName="rabbitmq" containerID="cri-o://e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366" gracePeriod=604795 Dec 09 17:31:49 crc kubenswrapper[4954]: I1209 17:31:49.357373 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 17:31:49 crc kubenswrapper[4954]: E1209 17:31:49.485379 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:31:49 crc kubenswrapper[4954]: E1209 17:31:49.485740 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:31:49 crc kubenswrapper[4954]: E1209 17:31:49.485927 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:31:50 crc kubenswrapper[4954]: I1209 17:31:50.137945 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="940c8621-fab6-4945-a800-aec88957e8cd" path="/var/lib/kubelet/pods/940c8621-fab6-4945-a800-aec88957e8cd/volumes" Dec 09 17:31:50 crc kubenswrapper[4954]: I1209 17:31:50.310718 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df","Type":"ContainerStarted","Data":"ee8b29d5458a49e26c406a99852737e2afac0b198dce1ea019ec80f0ce0572ff"} Dec 09 17:31:50 crc kubenswrapper[4954]: I1209 17:31:50.311128 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df","Type":"ContainerStarted","Data":"b678d0c82ca61b89b6c111d7e1efa8f932bfcd88a35401d638df101384086722"} Dec 09 17:31:50 crc kubenswrapper[4954]: I1209 17:31:50.483022 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="28718e77-f5b5-4d67-afbc-c1157354bc47" containerName="rabbitmq" containerID="cri-o://343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb" gracePeriod=604795 Dec 09 17:31:51 crc kubenswrapper[4954]: I1209 17:31:51.875784 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="28718e77-f5b5-4d67-afbc-c1157354bc47" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.126:5671: connect: connection refused" Dec 09 17:31:52 crc kubenswrapper[4954]: I1209 17:31:52.316109 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="3645ed14-3d97-496e-a683-2e56ea33fec9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.127:5671: connect: connection refused" Dec 09 17:31:52 crc kubenswrapper[4954]: I1209 17:31:52.331764 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df","Type":"ContainerStarted","Data":"d37b0f907708136ae7748985bbd4133309c83dcb6a257815114239956f503f6e"} Dec 09 17:31:53 crc kubenswrapper[4954]: E1209 17:31:53.284587 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:31:53 crc kubenswrapper[4954]: I1209 17:31:53.345907 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df","Type":"ContainerStarted","Data":"65c95e23f4678bdd60ccfb5c0958faea3bd3c1f5dcd7ce7e9f3051be2b642fd0"} Dec 09 17:31:53 crc kubenswrapper[4954]: I1209 17:31:53.346118 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 17:31:53 crc kubenswrapper[4954]: E1209 17:31:53.349017 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:31:54 crc kubenswrapper[4954]: E1209 17:31:54.358131 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.903466 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.981375 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-plugins\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.983052 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-config-data\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.983303 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-plugins-conf\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.983417 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3645ed14-3d97-496e-a683-2e56ea33fec9-erlang-cookie-secret\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.983586 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-confd\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.983653 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.983686 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-server-conf\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.983772 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-tls\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.983836 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmjsq\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-kube-api-access-dmjsq\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.983866 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3645ed14-3d97-496e-a683-2e56ea33fec9-pod-info\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.983998 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-erlang-cookie\") pod \"3645ed14-3d97-496e-a683-2e56ea33fec9\" (UID: \"3645ed14-3d97-496e-a683-2e56ea33fec9\") " Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.992069 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:31:55 crc kubenswrapper[4954]: I1209 17:31:55.996963 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3645ed14-3d97-496e-a683-2e56ea33fec9-pod-info" (OuterVolumeSpecName: "pod-info") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.000677 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.005016 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.014132 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3645ed14-3d97-496e-a683-2e56ea33fec9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.015851 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.022645 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-kube-api-access-dmjsq" (OuterVolumeSpecName: "kube-api-access-dmjsq") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "kube-api-access-dmjsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.034689 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.095068 4954 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3645ed14-3d97-496e-a683-2e56ea33fec9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.095814 4954 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.095830 4954 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.095842 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmjsq\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-kube-api-access-dmjsq\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.095865 4954 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3645ed14-3d97-496e-a683-2e56ea33fec9-pod-info\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.095878 4954 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.095888 4954 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.095896 4954 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.108730 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-config-data" (OuterVolumeSpecName: "config-data") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.125124 4954 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.151424 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-server-conf" (OuterVolumeSpecName: "server-conf") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.197488 4954 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.197529 4954 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-server-conf\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.197543 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3645ed14-3d97-496e-a683-2e56ea33fec9-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.231068 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3645ed14-3d97-496e-a683-2e56ea33fec9" (UID: "3645ed14-3d97-496e-a683-2e56ea33fec9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:31:56 crc kubenswrapper[4954]: E1209 17:31:56.264255 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:31:56 crc kubenswrapper[4954]: E1209 17:31:56.264332 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:31:56 crc kubenswrapper[4954]: E1209 17:31:56.264503 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:31:56 crc kubenswrapper[4954]: E1209 17:31:56.265693 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.300188 4954 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3645ed14-3d97-496e-a683-2e56ea33fec9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.388525 4954 generic.go:334] "Generic (PLEG): container finished" podID="3645ed14-3d97-496e-a683-2e56ea33fec9" containerID="e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366" exitCode=0 Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.388583 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3645ed14-3d97-496e-a683-2e56ea33fec9","Type":"ContainerDied","Data":"e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366"} Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.388605 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.388643 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3645ed14-3d97-496e-a683-2e56ea33fec9","Type":"ContainerDied","Data":"f09af2ede48c6c2cac1998e75e7a105a2e32d41d51df5ef6f04a13d9226440c0"} Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.388667 4954 scope.go:117] "RemoveContainer" containerID="e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.431132 4954 scope.go:117] "RemoveContainer" containerID="e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.487171 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.499868 4954 scope.go:117] "RemoveContainer" containerID="e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.507340 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 17:31:56 crc kubenswrapper[4954]: E1209 17:31:56.507776 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366\": container with ID starting with e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366 not found: ID does not exist" containerID="e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.507830 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366"} err="failed to get container status \"e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366\": rpc error: code = NotFound desc = could not find container \"e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366\": container with ID starting with e99754c6760b4442711c7728dd0358577d56959e64b8ee0264e51b4848103366 not found: ID does not exist" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.507865 4954 scope.go:117] "RemoveContainer" containerID="e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6" Dec 09 17:31:56 crc kubenswrapper[4954]: E1209 17:31:56.513791 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6\": container with ID starting with e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6 not found: ID does not exist" containerID="e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.513845 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6"} err="failed to get container status \"e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6\": rpc error: code = NotFound desc = could not find container \"e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6\": container with ID starting with e2d3686ceb1278f92256d895853531f8e7f7a5f45d56cd5598129467a61472d6 not found: ID does not exist" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.553020 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 17:31:56 crc kubenswrapper[4954]: E1209 17:31:56.553718 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3645ed14-3d97-496e-a683-2e56ea33fec9" containerName="rabbitmq" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.553738 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3645ed14-3d97-496e-a683-2e56ea33fec9" containerName="rabbitmq" Dec 09 17:31:56 crc kubenswrapper[4954]: E1209 17:31:56.553771 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3645ed14-3d97-496e-a683-2e56ea33fec9" containerName="setup-container" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.553780 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3645ed14-3d97-496e-a683-2e56ea33fec9" containerName="setup-container" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.554073 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="3645ed14-3d97-496e-a683-2e56ea33fec9" containerName="rabbitmq" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.555905 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.565979 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.572520 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.572746 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.572750 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.572838 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8smf5" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.572881 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.573121 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.646129 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712045 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ccb386f6-c327-4832-a461-7dd5cb0b44b7-config-data\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712161 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712230 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ccb386f6-c327-4832-a461-7dd5cb0b44b7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712282 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712331 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ccb386f6-c327-4832-a461-7dd5cb0b44b7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712376 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712491 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712552 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712578 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ccb386f6-c327-4832-a461-7dd5cb0b44b7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712614 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ccb386f6-c327-4832-a461-7dd5cb0b44b7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.712752 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lq2c\" (UniqueName: \"kubernetes.io/projected/ccb386f6-c327-4832-a461-7dd5cb0b44b7-kube-api-access-4lq2c\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.814738 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lq2c\" (UniqueName: \"kubernetes.io/projected/ccb386f6-c327-4832-a461-7dd5cb0b44b7-kube-api-access-4lq2c\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.815069 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ccb386f6-c327-4832-a461-7dd5cb0b44b7-config-data\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.815198 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.815297 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ccb386f6-c327-4832-a461-7dd5cb0b44b7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.815401 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.815487 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ccb386f6-c327-4832-a461-7dd5cb0b44b7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.815574 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.815779 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.815917 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.815990 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ccb386f6-c327-4832-a461-7dd5cb0b44b7-config-data\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.816011 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.816076 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.816009 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ccb386f6-c327-4832-a461-7dd5cb0b44b7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.816184 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ccb386f6-c327-4832-a461-7dd5cb0b44b7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.817002 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ccb386f6-c327-4832-a461-7dd5cb0b44b7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.823193 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ccb386f6-c327-4832-a461-7dd5cb0b44b7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.824549 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.825211 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ccb386f6-c327-4832-a461-7dd5cb0b44b7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.825343 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.828736 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ccb386f6-c327-4832-a461-7dd5cb0b44b7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.838297 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ccb386f6-c327-4832-a461-7dd5cb0b44b7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.855407 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lq2c\" (UniqueName: \"kubernetes.io/projected/ccb386f6-c327-4832-a461-7dd5cb0b44b7-kube-api-access-4lq2c\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:56 crc kubenswrapper[4954]: I1209 17:31:56.878146 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"ccb386f6-c327-4832-a461-7dd5cb0b44b7\") " pod="openstack/rabbitmq-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: E1209 17:31:57.033653 4954 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28718e77_f5b5_4d67_afbc_c1157354bc47.slice/crio-conmon-343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28718e77_f5b5_4d67_afbc_c1157354bc47.slice/crio-343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb.scope\": RecentStats: unable to find data in memory cache]" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.064234 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.276942 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.332708 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-plugins-conf\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.332896 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-server-conf\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.332955 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-tls\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.332989 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-erlang-cookie\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.333016 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-config-data\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.333104 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-plugins\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.333231 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-confd\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.333268 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.333299 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/28718e77-f5b5-4d67-afbc-c1157354bc47-pod-info\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.333355 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxtjv\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-kube-api-access-wxtjv\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.333516 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/28718e77-f5b5-4d67-afbc-c1157354bc47-erlang-cookie-secret\") pod \"28718e77-f5b5-4d67-afbc-c1157354bc47\" (UID: \"28718e77-f5b5-4d67-afbc-c1157354bc47\") " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.335466 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.336161 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.336263 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.342588 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28718e77-f5b5-4d67-afbc-c1157354bc47-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.342955 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/28718e77-f5b5-4d67-afbc-c1157354bc47-pod-info" (OuterVolumeSpecName: "pod-info") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.343205 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-kube-api-access-wxtjv" (OuterVolumeSpecName: "kube-api-access-wxtjv") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "kube-api-access-wxtjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.343388 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.347874 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.384556 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-config-data" (OuterVolumeSpecName: "config-data") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.431765 4954 generic.go:334] "Generic (PLEG): container finished" podID="28718e77-f5b5-4d67-afbc-c1157354bc47" containerID="343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb" exitCode=0 Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.431853 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"28718e77-f5b5-4d67-afbc-c1157354bc47","Type":"ContainerDied","Data":"343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb"} Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.431884 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"28718e77-f5b5-4d67-afbc-c1157354bc47","Type":"ContainerDied","Data":"c23918ecab3c55b33ef7105f7a0a5eba7386421da8b3660f44ecb96d397fb104"} Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.431904 4954 scope.go:117] "RemoveContainer" containerID="343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.432037 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.438857 4954 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.439031 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.439120 4954 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.440072 4954 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.440282 4954 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.440355 4954 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/28718e77-f5b5-4d67-afbc-c1157354bc47-pod-info\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.440411 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxtjv\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-kube-api-access-wxtjv\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.440476 4954 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/28718e77-f5b5-4d67-afbc-c1157354bc47-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.440533 4954 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.467734 4954 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.474062 4954 scope.go:117] "RemoveContainer" containerID="fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.475553 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-server-conf" (OuterVolumeSpecName: "server-conf") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.545512 4954 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/28718e77-f5b5-4d67-afbc-c1157354bc47-server-conf\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.545548 4954 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.552953 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "28718e77-f5b5-4d67-afbc-c1157354bc47" (UID: "28718e77-f5b5-4d67-afbc-c1157354bc47"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.647539 4954 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/28718e77-f5b5-4d67-afbc-c1157354bc47-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.675804 4954 scope.go:117] "RemoveContainer" containerID="343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb" Dec 09 17:31:57 crc kubenswrapper[4954]: E1209 17:31:57.676984 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb\": container with ID starting with 343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb not found: ID does not exist" containerID="343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.677041 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb"} err="failed to get container status \"343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb\": rpc error: code = NotFound desc = could not find container \"343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb\": container with ID starting with 343ed64294a1dfc068731dbcfbf84bf835ae720f57c33fb09ec7292e3e2da3bb not found: ID does not exist" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.677074 4954 scope.go:117] "RemoveContainer" containerID="fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad" Dec 09 17:31:57 crc kubenswrapper[4954]: E1209 17:31:57.681200 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad\": container with ID starting with fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad not found: ID does not exist" containerID="fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.681245 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad"} err="failed to get container status \"fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad\": rpc error: code = NotFound desc = could not find container \"fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad\": container with ID starting with fadfe68654a005e6563c51e2eee956b29ec52c8174b7bcf9e898261d234db6ad not found: ID does not exist" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.786126 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.842750 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.858886 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.871777 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 17:31:57 crc kubenswrapper[4954]: E1209 17:31:57.872331 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28718e77-f5b5-4d67-afbc-c1157354bc47" containerName="rabbitmq" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.872355 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="28718e77-f5b5-4d67-afbc-c1157354bc47" containerName="rabbitmq" Dec 09 17:31:57 crc kubenswrapper[4954]: E1209 17:31:57.872384 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28718e77-f5b5-4d67-afbc-c1157354bc47" containerName="setup-container" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.872393 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="28718e77-f5b5-4d67-afbc-c1157354bc47" containerName="setup-container" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.872704 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="28718e77-f5b5-4d67-afbc-c1157354bc47" containerName="rabbitmq" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.874145 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.879212 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.879362 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.879529 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.879733 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.879760 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.879952 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-x79kt" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.881984 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.886752 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958094 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958191 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fa9d388c-71e6-4332-9e32-de049db360e1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958267 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fa9d388c-71e6-4332-9e32-de049db360e1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958287 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958320 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958340 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fa9d388c-71e6-4332-9e32-de049db360e1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958404 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fa9d388c-71e6-4332-9e32-de049db360e1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958447 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958510 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958567 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdvq6\" (UniqueName: \"kubernetes.io/projected/fa9d388c-71e6-4332-9e32-de049db360e1-kube-api-access-xdvq6\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:57 crc kubenswrapper[4954]: I1209 17:31:57.958625 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa9d388c-71e6-4332-9e32-de049db360e1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.061115 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062073 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdvq6\" (UniqueName: \"kubernetes.io/projected/fa9d388c-71e6-4332-9e32-de049db360e1-kube-api-access-xdvq6\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062140 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa9d388c-71e6-4332-9e32-de049db360e1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062186 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062248 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fa9d388c-71e6-4332-9e32-de049db360e1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062321 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fa9d388c-71e6-4332-9e32-de049db360e1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062341 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062387 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062418 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fa9d388c-71e6-4332-9e32-de049db360e1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062501 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fa9d388c-71e6-4332-9e32-de049db360e1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062571 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.062850 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.063015 4954 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.063327 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa9d388c-71e6-4332-9e32-de049db360e1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.066478 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fa9d388c-71e6-4332-9e32-de049db360e1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.069100 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fa9d388c-71e6-4332-9e32-de049db360e1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.069958 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.071079 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.071173 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fa9d388c-71e6-4332-9e32-de049db360e1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.071859 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fa9d388c-71e6-4332-9e32-de049db360e1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.076323 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fa9d388c-71e6-4332-9e32-de049db360e1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.087442 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdvq6\" (UniqueName: \"kubernetes.io/projected/fa9d388c-71e6-4332-9e32-de049db360e1-kube-api-access-xdvq6\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.109038 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fa9d388c-71e6-4332-9e32-de049db360e1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.139625 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28718e77-f5b5-4d67-afbc-c1157354bc47" path="/var/lib/kubelet/pods/28718e77-f5b5-4d67-afbc-c1157354bc47/volumes" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.141802 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3645ed14-3d97-496e-a683-2e56ea33fec9" path="/var/lib/kubelet/pods/3645ed14-3d97-496e-a683-2e56ea33fec9/volumes" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.212272 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.458941 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ccb386f6-c327-4832-a461-7dd5cb0b44b7","Type":"ContainerStarted","Data":"f32f7b75c30d11ad502495695a35204df8d4042135eb1ffa67a65faae41eff23"} Dec 09 17:31:58 crc kubenswrapper[4954]: I1209 17:31:58.799682 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 17:31:59 crc kubenswrapper[4954]: I1209 17:31:59.474186 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fa9d388c-71e6-4332-9e32-de049db360e1","Type":"ContainerStarted","Data":"49c1722d5f9d0165524323c2244470a5f67ed248c53965adfc208c63d33681a3"} Dec 09 17:32:00 crc kubenswrapper[4954]: I1209 17:32:00.488914 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ccb386f6-c327-4832-a461-7dd5cb0b44b7","Type":"ContainerStarted","Data":"a32f05bb2432bbdf2b1c3b6409654503c423a884018f954d15727101b0c2d600"} Dec 09 17:32:01 crc kubenswrapper[4954]: I1209 17:32:01.521619 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fa9d388c-71e6-4332-9e32-de049db360e1","Type":"ContainerStarted","Data":"b6782ba8c5d7afd9c368e32c54b29a8b92d5f6fa19d32fc6de8540ceaeab2b06"} Dec 09 17:32:01 crc kubenswrapper[4954]: I1209 17:32:01.968327 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-t27cq"] Dec 09 17:32:01 crc kubenswrapper[4954]: I1209 17:32:01.971912 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:01 crc kubenswrapper[4954]: I1209 17:32:01.976771 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.000946 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-t27cq"] Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.074524 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-config\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.074581 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql4w5\" (UniqueName: \"kubernetes.io/projected/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-kube-api-access-ql4w5\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.074620 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.074842 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.074995 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.075039 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.075270 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.177580 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-config\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.177666 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql4w5\" (UniqueName: \"kubernetes.io/projected/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-kube-api-access-ql4w5\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.177711 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.177811 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.177900 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.177917 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.178087 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.178866 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.178920 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.178917 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.179009 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.179153 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.179551 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-config\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.205679 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql4w5\" (UniqueName: \"kubernetes.io/projected/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-kube-api-access-ql4w5\") pod \"dnsmasq-dns-7d84b4d45c-t27cq\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:02 crc kubenswrapper[4954]: I1209 17:32:02.322960 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:03 crc kubenswrapper[4954]: I1209 17:32:03.180324 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-t27cq"] Dec 09 17:32:03 crc kubenswrapper[4954]: W1209 17:32:03.191853 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c74f16c_2fb6_4f7e_a2fe_ec7bf5035b7d.slice/crio-c08bd5bdfc6ab0de05f53713a5ead71732c6f8cad3f6d4a0b33c41dc19cacb51 WatchSource:0}: Error finding container c08bd5bdfc6ab0de05f53713a5ead71732c6f8cad3f6d4a0b33c41dc19cacb51: Status 404 returned error can't find the container with id c08bd5bdfc6ab0de05f53713a5ead71732c6f8cad3f6d4a0b33c41dc19cacb51 Dec 09 17:32:03 crc kubenswrapper[4954]: I1209 17:32:03.567171 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" event={"ID":"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d","Type":"ContainerStarted","Data":"ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983"} Dec 09 17:32:03 crc kubenswrapper[4954]: I1209 17:32:03.567552 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" event={"ID":"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d","Type":"ContainerStarted","Data":"c08bd5bdfc6ab0de05f53713a5ead71732c6f8cad3f6d4a0b33c41dc19cacb51"} Dec 09 17:32:04 crc kubenswrapper[4954]: I1209 17:32:04.582387 4954 generic.go:334] "Generic (PLEG): container finished" podID="9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" containerID="ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983" exitCode=0 Dec 09 17:32:04 crc kubenswrapper[4954]: I1209 17:32:04.582722 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" event={"ID":"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d","Type":"ContainerDied","Data":"ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983"} Dec 09 17:32:04 crc kubenswrapper[4954]: I1209 17:32:04.582756 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" event={"ID":"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d","Type":"ContainerStarted","Data":"4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148"} Dec 09 17:32:04 crc kubenswrapper[4954]: I1209 17:32:04.584884 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:04 crc kubenswrapper[4954]: I1209 17:32:04.613676 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" podStartSLOduration=3.613643284 podStartE2EDuration="3.613643284s" podCreationTimestamp="2025-12-09 17:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:32:04.608736321 +0000 UTC m=+2120.996910151" watchObservedRunningTime="2025-12-09 17:32:04.613643284 +0000 UTC m=+2121.001817114" Dec 09 17:32:06 crc kubenswrapper[4954]: I1209 17:32:06.134085 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 17:32:06 crc kubenswrapper[4954]: E1209 17:32:06.232961 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:32:06 crc kubenswrapper[4954]: E1209 17:32:06.233034 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:32:06 crc kubenswrapper[4954]: E1209 17:32:06.233245 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:32:06 crc kubenswrapper[4954]: E1209 17:32:06.235149 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:32:06 crc kubenswrapper[4954]: E1209 17:32:06.612562 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:32:09 crc kubenswrapper[4954]: E1209 17:32:09.122656 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.324757 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.401953 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl"] Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.402307 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" podUID="507b400c-fcae-4257-a6e5-2485da532750" containerName="dnsmasq-dns" containerID="cri-o://13799642b80ec8b052de87ce60ecb3f5d5f44d5347698e50b625609de5897637" gracePeriod=10 Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.647675 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-v6xcs"] Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.650684 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.671227 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-v6xcs"] Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.748246 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-dns-svc\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.748734 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-ovsdbserver-sb\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.748910 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-ovsdbserver-nb\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.749021 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-dns-swift-storage-0\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.749134 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-openstack-edpm-ipam\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.749299 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-config\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.749447 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w7qd\" (UniqueName: \"kubernetes.io/projected/c37d726c-59ef-4dd7-8a00-32417dfe1a19-kube-api-access-5w7qd\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.750532 4954 generic.go:334] "Generic (PLEG): container finished" podID="507b400c-fcae-4257-a6e5-2485da532750" containerID="13799642b80ec8b052de87ce60ecb3f5d5f44d5347698e50b625609de5897637" exitCode=0 Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.750583 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" event={"ID":"507b400c-fcae-4257-a6e5-2485da532750","Type":"ContainerDied","Data":"13799642b80ec8b052de87ce60ecb3f5d5f44d5347698e50b625609de5897637"} Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.851721 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-ovsdbserver-nb\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.851786 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-dns-swift-storage-0\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.851853 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-openstack-edpm-ipam\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.851910 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-config\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.851945 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w7qd\" (UniqueName: \"kubernetes.io/projected/c37d726c-59ef-4dd7-8a00-32417dfe1a19-kube-api-access-5w7qd\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.852030 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-dns-svc\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.852125 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-ovsdbserver-sb\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.853119 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-ovsdbserver-sb\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.853477 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-openstack-edpm-ipam\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.854280 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-dns-swift-storage-0\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.854515 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-dns-svc\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.854890 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-ovsdbserver-nb\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.855299 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c37d726c-59ef-4dd7-8a00-32417dfe1a19-config\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:12 crc kubenswrapper[4954]: I1209 17:32:12.894178 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w7qd\" (UniqueName: \"kubernetes.io/projected/c37d726c-59ef-4dd7-8a00-32417dfe1a19-kube-api-access-5w7qd\") pod \"dnsmasq-dns-6559847fc9-v6xcs\" (UID: \"c37d726c-59ef-4dd7-8a00-32417dfe1a19\") " pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:12.998095 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.242587 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.286891 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-swift-storage-0\") pod \"507b400c-fcae-4257-a6e5-2485da532750\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.287859 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c86ql\" (UniqueName: \"kubernetes.io/projected/507b400c-fcae-4257-a6e5-2485da532750-kube-api-access-c86ql\") pod \"507b400c-fcae-4257-a6e5-2485da532750\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.287922 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-svc\") pod \"507b400c-fcae-4257-a6e5-2485da532750\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.287969 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-sb\") pod \"507b400c-fcae-4257-a6e5-2485da532750\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.288046 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-config\") pod \"507b400c-fcae-4257-a6e5-2485da532750\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.288150 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-nb\") pod \"507b400c-fcae-4257-a6e5-2485da532750\" (UID: \"507b400c-fcae-4257-a6e5-2485da532750\") " Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.293491 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/507b400c-fcae-4257-a6e5-2485da532750-kube-api-access-c86ql" (OuterVolumeSpecName: "kube-api-access-c86ql") pod "507b400c-fcae-4257-a6e5-2485da532750" (UID: "507b400c-fcae-4257-a6e5-2485da532750"). InnerVolumeSpecName "kube-api-access-c86ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.349540 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "507b400c-fcae-4257-a6e5-2485da532750" (UID: "507b400c-fcae-4257-a6e5-2485da532750"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.349636 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "507b400c-fcae-4257-a6e5-2485da532750" (UID: "507b400c-fcae-4257-a6e5-2485da532750"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.357579 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "507b400c-fcae-4257-a6e5-2485da532750" (UID: "507b400c-fcae-4257-a6e5-2485da532750"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.373215 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "507b400c-fcae-4257-a6e5-2485da532750" (UID: "507b400c-fcae-4257-a6e5-2485da532750"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.380416 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-config" (OuterVolumeSpecName: "config") pod "507b400c-fcae-4257-a6e5-2485da532750" (UID: "507b400c-fcae-4257-a6e5-2485da532750"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.392886 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.392921 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c86ql\" (UniqueName: \"kubernetes.io/projected/507b400c-fcae-4257-a6e5-2485da532750-kube-api-access-c86ql\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.392934 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.392945 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.392954 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.392963 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/507b400c-fcae-4257-a6e5-2485da532750-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.573497 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-v6xcs"] Dec 09 17:32:13 crc kubenswrapper[4954]: W1209 17:32:13.582416 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc37d726c_59ef_4dd7_8a00_32417dfe1a19.slice/crio-ea2aee44ae36b1210b494d35c81536057a668fc9970da9eed86b69fb890b2990 WatchSource:0}: Error finding container ea2aee44ae36b1210b494d35c81536057a668fc9970da9eed86b69fb890b2990: Status 404 returned error can't find the container with id ea2aee44ae36b1210b494d35c81536057a668fc9970da9eed86b69fb890b2990 Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.755741 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.755799 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.764911 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" event={"ID":"c37d726c-59ef-4dd7-8a00-32417dfe1a19","Type":"ContainerStarted","Data":"ea2aee44ae36b1210b494d35c81536057a668fc9970da9eed86b69fb890b2990"} Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.767705 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" event={"ID":"507b400c-fcae-4257-a6e5-2485da532750","Type":"ContainerDied","Data":"851aa647fc7d9646eb4f96017a9dc0e25994139d8f96a9121a32c64a4b43d762"} Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.767749 4954 scope.go:117] "RemoveContainer" containerID="13799642b80ec8b052de87ce60ecb3f5d5f44d5347698e50b625609de5897637" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.767909 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.797003 4954 scope.go:117] "RemoveContainer" containerID="e3fa3ea2965b62e667fd670ca8dcfedf5079b9846de4579248fab4a0b8660b7a" Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.822647 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl"] Dec 09 17:32:13 crc kubenswrapper[4954]: I1209 17:32:13.833820 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-kw4bl"] Dec 09 17:32:14 crc kubenswrapper[4954]: I1209 17:32:14.160953 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="507b400c-fcae-4257-a6e5-2485da532750" path="/var/lib/kubelet/pods/507b400c-fcae-4257-a6e5-2485da532750/volumes" Dec 09 17:32:14 crc kubenswrapper[4954]: I1209 17:32:14.780760 4954 generic.go:334] "Generic (PLEG): container finished" podID="c37d726c-59ef-4dd7-8a00-32417dfe1a19" containerID="0b65686504ca65a4eadeceaa7999f264335a91cc520fe5727b8aac0c7be504a9" exitCode=0 Dec 09 17:32:14 crc kubenswrapper[4954]: I1209 17:32:14.780943 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" event={"ID":"c37d726c-59ef-4dd7-8a00-32417dfe1a19","Type":"ContainerDied","Data":"0b65686504ca65a4eadeceaa7999f264335a91cc520fe5727b8aac0c7be504a9"} Dec 09 17:32:15 crc kubenswrapper[4954]: I1209 17:32:15.798367 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" event={"ID":"c37d726c-59ef-4dd7-8a00-32417dfe1a19","Type":"ContainerStarted","Data":"4fa8c95c3e8046ac694156ee9ec03bb4373e69770ab8a21d118f4d0e0b554910"} Dec 09 17:32:15 crc kubenswrapper[4954]: I1209 17:32:15.799013 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:15 crc kubenswrapper[4954]: I1209 17:32:15.831865 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" podStartSLOduration=3.831837456 podStartE2EDuration="3.831837456s" podCreationTimestamp="2025-12-09 17:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:32:15.81952851 +0000 UTC m=+2132.207702340" watchObservedRunningTime="2025-12-09 17:32:15.831837456 +0000 UTC m=+2132.220011266" Dec 09 17:32:19 crc kubenswrapper[4954]: E1209 17:32:19.123627 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:22.999786 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6559847fc9-v6xcs" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.085263 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-t27cq"] Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.085782 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" podUID="9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" containerName="dnsmasq-dns" containerID="cri-o://4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148" gracePeriod=10 Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.697165 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.751333 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-svc\") pod \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.751405 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-sb\") pod \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.751547 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-nb\") pod \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.751603 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-swift-storage-0\") pod \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.751635 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-openstack-edpm-ipam\") pod \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.751662 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-config\") pod \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.751899 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql4w5\" (UniqueName: \"kubernetes.io/projected/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-kube-api-access-ql4w5\") pod \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\" (UID: \"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d\") " Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.793799 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-kube-api-access-ql4w5" (OuterVolumeSpecName: "kube-api-access-ql4w5") pod "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" (UID: "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d"). InnerVolumeSpecName "kube-api-access-ql4w5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.854246 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql4w5\" (UniqueName: \"kubernetes.io/projected/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-kube-api-access-ql4w5\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.906115 4954 generic.go:334] "Generic (PLEG): container finished" podID="9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" containerID="4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148" exitCode=0 Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.906165 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" event={"ID":"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d","Type":"ContainerDied","Data":"4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148"} Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.906194 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" event={"ID":"9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d","Type":"ContainerDied","Data":"c08bd5bdfc6ab0de05f53713a5ead71732c6f8cad3f6d4a0b33c41dc19cacb51"} Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.906210 4954 scope.go:117] "RemoveContainer" containerID="4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.906381 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-t27cq" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.907124 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" (UID: "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.916462 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" (UID: "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.956257 4954 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.956300 4954 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.964181 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" (UID: "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.988537 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" (UID: "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:23 crc kubenswrapper[4954]: I1209 17:32:23.991059 4954 scope.go:117] "RemoveContainer" containerID="ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.031335 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" (UID: "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.034089 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-config" (OuterVolumeSpecName: "config") pod "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" (UID: "9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.057843 4954 scope.go:117] "RemoveContainer" containerID="4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.058515 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.058564 4954 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.058578 4954 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.058605 4954 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d-config\") on node \"crc\" DevicePath \"\"" Dec 09 17:32:24 crc kubenswrapper[4954]: E1209 17:32:24.063773 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148\": container with ID starting with 4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148 not found: ID does not exist" containerID="4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.063829 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148"} err="failed to get container status \"4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148\": rpc error: code = NotFound desc = could not find container \"4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148\": container with ID starting with 4ed4a4e27a0a36f6bd2c0f998064f149a0fbcc35d77f122bbcc7b105aefcb148 not found: ID does not exist" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.063858 4954 scope.go:117] "RemoveContainer" containerID="ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983" Dec 09 17:32:24 crc kubenswrapper[4954]: E1209 17:32:24.068495 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983\": container with ID starting with ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983 not found: ID does not exist" containerID="ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.068566 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983"} err="failed to get container status \"ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983\": rpc error: code = NotFound desc = could not find container \"ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983\": container with ID starting with ededf348410264ace7c325bbcdb8e28cafefb68ac96efc55ece203207a12e983 not found: ID does not exist" Dec 09 17:32:24 crc kubenswrapper[4954]: E1209 17:32:24.260939 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:32:24 crc kubenswrapper[4954]: E1209 17:32:24.261292 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:32:24 crc kubenswrapper[4954]: E1209 17:32:24.261450 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:32:24 crc kubenswrapper[4954]: E1209 17:32:24.265507 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.288137 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-t27cq"] Dec 09 17:32:24 crc kubenswrapper[4954]: I1209 17:32:24.298654 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-t27cq"] Dec 09 17:32:26 crc kubenswrapper[4954]: I1209 17:32:26.133381 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" path="/var/lib/kubelet/pods/9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d/volumes" Dec 09 17:32:30 crc kubenswrapper[4954]: E1209 17:32:30.245806 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:32:30 crc kubenswrapper[4954]: E1209 17:32:30.246535 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:32:30 crc kubenswrapper[4954]: E1209 17:32:30.246801 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:32:30 crc kubenswrapper[4954]: E1209 17:32:30.248012 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:32:33 crc kubenswrapper[4954]: I1209 17:32:33.011723 4954 generic.go:334] "Generic (PLEG): container finished" podID="ccb386f6-c327-4832-a461-7dd5cb0b44b7" containerID="a32f05bb2432bbdf2b1c3b6409654503c423a884018f954d15727101b0c2d600" exitCode=0 Dec 09 17:32:33 crc kubenswrapper[4954]: I1209 17:32:33.011817 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ccb386f6-c327-4832-a461-7dd5cb0b44b7","Type":"ContainerDied","Data":"a32f05bb2432bbdf2b1c3b6409654503c423a884018f954d15727101b0c2d600"} Dec 09 17:32:33 crc kubenswrapper[4954]: I1209 17:32:33.017171 4954 generic.go:334] "Generic (PLEG): container finished" podID="fa9d388c-71e6-4332-9e32-de049db360e1" containerID="b6782ba8c5d7afd9c368e32c54b29a8b92d5f6fa19d32fc6de8540ceaeab2b06" exitCode=0 Dec 09 17:32:33 crc kubenswrapper[4954]: I1209 17:32:33.017221 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fa9d388c-71e6-4332-9e32-de049db360e1","Type":"ContainerDied","Data":"b6782ba8c5d7afd9c368e32c54b29a8b92d5f6fa19d32fc6de8540ceaeab2b06"} Dec 09 17:32:34 crc kubenswrapper[4954]: I1209 17:32:34.032411 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ccb386f6-c327-4832-a461-7dd5cb0b44b7","Type":"ContainerStarted","Data":"909c1846ae3629960ec72d155ac502b70da57d17f22cfe6fe7851efb02ccf2b9"} Dec 09 17:32:34 crc kubenswrapper[4954]: I1209 17:32:34.033916 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 09 17:32:34 crc kubenswrapper[4954]: I1209 17:32:34.036643 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fa9d388c-71e6-4332-9e32-de049db360e1","Type":"ContainerStarted","Data":"f4ea449986a1abda2b5477bd54bb076361a0244213e928a91015a44a2654423e"} Dec 09 17:32:34 crc kubenswrapper[4954]: I1209 17:32:34.036942 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:32:34 crc kubenswrapper[4954]: I1209 17:32:34.100619 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.100565696 podStartE2EDuration="37.100565696s" podCreationTimestamp="2025-12-09 17:31:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:32:34.097403707 +0000 UTC m=+2150.485577527" watchObservedRunningTime="2025-12-09 17:32:34.100565696 +0000 UTC m=+2150.488739526" Dec 09 17:32:34 crc kubenswrapper[4954]: I1209 17:32:34.108285 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.108256468 podStartE2EDuration="38.108256468s" podCreationTimestamp="2025-12-09 17:31:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 17:32:34.070494975 +0000 UTC m=+2150.458668795" watchObservedRunningTime="2025-12-09 17:32:34.108256468 +0000 UTC m=+2150.496430298" Dec 09 17:32:36 crc kubenswrapper[4954]: E1209 17:32:36.122853 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.141118 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s"] Dec 09 17:32:36 crc kubenswrapper[4954]: E1209 17:32:36.148727 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="507b400c-fcae-4257-a6e5-2485da532750" containerName="dnsmasq-dns" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.148841 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="507b400c-fcae-4257-a6e5-2485da532750" containerName="dnsmasq-dns" Dec 09 17:32:36 crc kubenswrapper[4954]: E1209 17:32:36.149037 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" containerName="dnsmasq-dns" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.149109 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" containerName="dnsmasq-dns" Dec 09 17:32:36 crc kubenswrapper[4954]: E1209 17:32:36.149192 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" containerName="init" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.149248 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" containerName="init" Dec 09 17:32:36 crc kubenswrapper[4954]: E1209 17:32:36.149318 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="507b400c-fcae-4257-a6e5-2485da532750" containerName="init" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.149393 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="507b400c-fcae-4257-a6e5-2485da532750" containerName="init" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.149734 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c74f16c-2fb6-4f7e-a2fe-ec7bf5035b7d" containerName="dnsmasq-dns" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.149839 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="507b400c-fcae-4257-a6e5-2485da532750" containerName="dnsmasq-dns" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.150778 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.167525 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s"] Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.194534 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.194764 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.194789 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.194948 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.296933 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.297392 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7vb8\" (UniqueName: \"kubernetes.io/projected/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-kube-api-access-t7vb8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.297746 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.297944 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.405417 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.405976 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7vb8\" (UniqueName: \"kubernetes.io/projected/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-kube-api-access-t7vb8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.406013 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.406066 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.415573 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.417310 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.417953 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.427400 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7vb8\" (UniqueName: \"kubernetes.io/projected/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-kube-api-access-t7vb8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:36 crc kubenswrapper[4954]: I1209 17:32:36.519735 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:32:37 crc kubenswrapper[4954]: I1209 17:32:37.219925 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s"] Dec 09 17:32:38 crc kubenswrapper[4954]: I1209 17:32:38.088443 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" event={"ID":"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6","Type":"ContainerStarted","Data":"bccea4e9430742a5693669a78946b47d621efc87a7b1e3fde89abcc595827ace"} Dec 09 17:32:41 crc kubenswrapper[4954]: E1209 17:32:41.122985 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:32:43 crc kubenswrapper[4954]: I1209 17:32:43.755578 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:32:43 crc kubenswrapper[4954]: I1209 17:32:43.755880 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:32:46 crc kubenswrapper[4954]: I1209 17:32:46.635951 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:32:47 crc kubenswrapper[4954]: I1209 17:32:47.068028 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 09 17:32:47 crc kubenswrapper[4954]: I1209 17:32:47.220145 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" event={"ID":"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6","Type":"ContainerStarted","Data":"c8ffceede183e2c966d82d7be25b3c3048f97644364007151a3e33167a701a64"} Dec 09 17:32:47 crc kubenswrapper[4954]: I1209 17:32:47.250814 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" podStartSLOduration=1.8402935089999999 podStartE2EDuration="11.25079084s" podCreationTimestamp="2025-12-09 17:32:36 +0000 UTC" firstStartedPulling="2025-12-09 17:32:37.222444321 +0000 UTC m=+2153.610618141" lastFinishedPulling="2025-12-09 17:32:46.632941642 +0000 UTC m=+2163.021115472" observedRunningTime="2025-12-09 17:32:47.244295747 +0000 UTC m=+2163.632469557" watchObservedRunningTime="2025-12-09 17:32:47.25079084 +0000 UTC m=+2163.638964660" Dec 09 17:32:48 crc kubenswrapper[4954]: E1209 17:32:48.125346 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:32:48 crc kubenswrapper[4954]: I1209 17:32:48.218669 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 09 17:32:55 crc kubenswrapper[4954]: E1209 17:32:55.122482 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:33:00 crc kubenswrapper[4954]: I1209 17:33:00.408182 4954 generic.go:334] "Generic (PLEG): container finished" podID="fc8a05c8-a7d1-440d-b800-f0a4217ec7f6" containerID="c8ffceede183e2c966d82d7be25b3c3048f97644364007151a3e33167a701a64" exitCode=0 Dec 09 17:33:00 crc kubenswrapper[4954]: I1209 17:33:00.409004 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" event={"ID":"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6","Type":"ContainerDied","Data":"c8ffceede183e2c966d82d7be25b3c3048f97644364007151a3e33167a701a64"} Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.112188 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6drgz"] Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.115393 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.122926 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcf8f\" (UniqueName: \"kubernetes.io/projected/d214bb4b-5cf1-4d40-9770-01d7bf789052-kube-api-access-bcf8f\") pod \"redhat-marketplace-6drgz\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.123832 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-catalog-content\") pod \"redhat-marketplace-6drgz\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.123960 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-utilities\") pod \"redhat-marketplace-6drgz\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.138226 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6drgz"] Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.227320 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-catalog-content\") pod \"redhat-marketplace-6drgz\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.227374 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-utilities\") pod \"redhat-marketplace-6drgz\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.227529 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcf8f\" (UniqueName: \"kubernetes.io/projected/d214bb4b-5cf1-4d40-9770-01d7bf789052-kube-api-access-bcf8f\") pod \"redhat-marketplace-6drgz\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.227905 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-catalog-content\") pod \"redhat-marketplace-6drgz\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.228157 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-utilities\") pod \"redhat-marketplace-6drgz\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.260181 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcf8f\" (UniqueName: \"kubernetes.io/projected/d214bb4b-5cf1-4d40-9770-01d7bf789052-kube-api-access-bcf8f\") pod \"redhat-marketplace-6drgz\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:01 crc kubenswrapper[4954]: I1209 17:33:01.445063 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.000258 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.048448 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7vb8\" (UniqueName: \"kubernetes.io/projected/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-kube-api-access-t7vb8\") pod \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.048518 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-repo-setup-combined-ca-bundle\") pod \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.048550 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-inventory\") pod \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.048732 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-ssh-key\") pod \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\" (UID: \"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6\") " Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.056842 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "fc8a05c8-a7d1-440d-b800-f0a4217ec7f6" (UID: "fc8a05c8-a7d1-440d-b800-f0a4217ec7f6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.056915 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-kube-api-access-t7vb8" (OuterVolumeSpecName: "kube-api-access-t7vb8") pod "fc8a05c8-a7d1-440d-b800-f0a4217ec7f6" (UID: "fc8a05c8-a7d1-440d-b800-f0a4217ec7f6"). InnerVolumeSpecName "kube-api-access-t7vb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.065327 4954 scope.go:117] "RemoveContainer" containerID="35b1cdfbacb3a5a691cd62bbe33d3dc479c80cc894a88016c2cf307a430ecd18" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.095564 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fc8a05c8-a7d1-440d-b800-f0a4217ec7f6" (UID: "fc8a05c8-a7d1-440d-b800-f0a4217ec7f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.111885 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-inventory" (OuterVolumeSpecName: "inventory") pod "fc8a05c8-a7d1-440d-b800-f0a4217ec7f6" (UID: "fc8a05c8-a7d1-440d-b800-f0a4217ec7f6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.111911 4954 scope.go:117] "RemoveContainer" containerID="7288584ac020f9f057251034402d4eeca1bba70fe00639b87d8c6bac614ed61e" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.117214 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6drgz"] Dec 09 17:33:02 crc kubenswrapper[4954]: W1209 17:33:02.121303 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd214bb4b_5cf1_4d40_9770_01d7bf789052.slice/crio-cbb25fd76f8ffd7373842447a4d2b5681f09b41c32991adc6a4020960117312f WatchSource:0}: Error finding container cbb25fd76f8ffd7373842447a4d2b5681f09b41c32991adc6a4020960117312f: Status 404 returned error can't find the container with id cbb25fd76f8ffd7373842447a4d2b5681f09b41c32991adc6a4020960117312f Dec 09 17:33:02 crc kubenswrapper[4954]: E1209 17:33:02.122381 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.151643 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7vb8\" (UniqueName: \"kubernetes.io/projected/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-kube-api-access-t7vb8\") on node \"crc\" DevicePath \"\"" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.151672 4954 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.151683 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.151694 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8a05c8-a7d1-440d-b800-f0a4217ec7f6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.433945 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" event={"ID":"fc8a05c8-a7d1-440d-b800-f0a4217ec7f6","Type":"ContainerDied","Data":"bccea4e9430742a5693669a78946b47d621efc87a7b1e3fde89abcc595827ace"} Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.434390 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bccea4e9430742a5693669a78946b47d621efc87a7b1e3fde89abcc595827ace" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.433973 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.435741 4954 generic.go:334] "Generic (PLEG): container finished" podID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerID="f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c" exitCode=0 Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.435781 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6drgz" event={"ID":"d214bb4b-5cf1-4d40-9770-01d7bf789052","Type":"ContainerDied","Data":"f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c"} Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.435810 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6drgz" event={"ID":"d214bb4b-5cf1-4d40-9770-01d7bf789052","Type":"ContainerStarted","Data":"cbb25fd76f8ffd7373842447a4d2b5681f09b41c32991adc6a4020960117312f"} Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.573677 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd"] Dec 09 17:33:02 crc kubenswrapper[4954]: E1209 17:33:02.574484 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc8a05c8-a7d1-440d-b800-f0a4217ec7f6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.574509 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc8a05c8-a7d1-440d-b800-f0a4217ec7f6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.574843 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc8a05c8-a7d1-440d-b800-f0a4217ec7f6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.576012 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.580517 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.580963 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.581122 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.581266 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.600402 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd"] Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.665903 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.666179 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p5lt\" (UniqueName: \"kubernetes.io/projected/43bffa1b-06d5-470d-b410-82f6acf14286-kube-api-access-2p5lt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.666286 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.666673 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.768810 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p5lt\" (UniqueName: \"kubernetes.io/projected/43bffa1b-06d5-470d-b410-82f6acf14286-kube-api-access-2p5lt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.768951 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.769041 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.769145 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.777313 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.780587 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.782740 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.791273 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p5lt\" (UniqueName: \"kubernetes.io/projected/43bffa1b-06d5-470d-b410-82f6acf14286-kube-api-access-2p5lt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:02 crc kubenswrapper[4954]: I1209 17:33:02.906909 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:33:03 crc kubenswrapper[4954]: I1209 17:33:03.450983 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6drgz" event={"ID":"d214bb4b-5cf1-4d40-9770-01d7bf789052","Type":"ContainerStarted","Data":"6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3"} Dec 09 17:33:03 crc kubenswrapper[4954]: I1209 17:33:03.570307 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd"] Dec 09 17:33:04 crc kubenswrapper[4954]: I1209 17:33:04.464495 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" event={"ID":"43bffa1b-06d5-470d-b410-82f6acf14286","Type":"ContainerStarted","Data":"0e4a7881f2c239a6fa3573ee2ef18170972f9b96f504ed856cf328a6cdf2d68d"} Dec 09 17:33:04 crc kubenswrapper[4954]: I1209 17:33:04.465093 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" event={"ID":"43bffa1b-06d5-470d-b410-82f6acf14286","Type":"ContainerStarted","Data":"155c4611b6d7c9e35ec6f179430532289dff8153b16852a92da0706129ab61bc"} Dec 09 17:33:04 crc kubenswrapper[4954]: I1209 17:33:04.466705 4954 generic.go:334] "Generic (PLEG): container finished" podID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerID="6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3" exitCode=0 Dec 09 17:33:04 crc kubenswrapper[4954]: I1209 17:33:04.466760 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6drgz" event={"ID":"d214bb4b-5cf1-4d40-9770-01d7bf789052","Type":"ContainerDied","Data":"6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3"} Dec 09 17:33:04 crc kubenswrapper[4954]: I1209 17:33:04.493895 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" podStartSLOduration=1.9862082490000001 podStartE2EDuration="2.493871519s" podCreationTimestamp="2025-12-09 17:33:02 +0000 UTC" firstStartedPulling="2025-12-09 17:33:03.553614039 +0000 UTC m=+2179.941787849" lastFinishedPulling="2025-12-09 17:33:04.061277289 +0000 UTC m=+2180.449451119" observedRunningTime="2025-12-09 17:33:04.49163992 +0000 UTC m=+2180.879813760" watchObservedRunningTime="2025-12-09 17:33:04.493871519 +0000 UTC m=+2180.882045339" Dec 09 17:33:06 crc kubenswrapper[4954]: I1209 17:33:06.495255 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6drgz" event={"ID":"d214bb4b-5cf1-4d40-9770-01d7bf789052","Type":"ContainerStarted","Data":"900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5"} Dec 09 17:33:06 crc kubenswrapper[4954]: I1209 17:33:06.519105 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6drgz" podStartSLOduration=2.582219945 podStartE2EDuration="5.519075689s" podCreationTimestamp="2025-12-09 17:33:01 +0000 UTC" firstStartedPulling="2025-12-09 17:33:02.438244558 +0000 UTC m=+2178.826418378" lastFinishedPulling="2025-12-09 17:33:05.375100302 +0000 UTC m=+2181.763274122" observedRunningTime="2025-12-09 17:33:06.511887193 +0000 UTC m=+2182.900061023" watchObservedRunningTime="2025-12-09 17:33:06.519075689 +0000 UTC m=+2182.907249509" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.494300 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n6ghn"] Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.499322 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.506340 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6ghn"] Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.611715 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mr4q\" (UniqueName: \"kubernetes.io/projected/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-kube-api-access-4mr4q\") pod \"certified-operators-n6ghn\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.611824 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-catalog-content\") pod \"certified-operators-n6ghn\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.611891 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-utilities\") pod \"certified-operators-n6ghn\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.713868 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-catalog-content\") pod \"certified-operators-n6ghn\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.713969 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-utilities\") pod \"certified-operators-n6ghn\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.714092 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mr4q\" (UniqueName: \"kubernetes.io/projected/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-kube-api-access-4mr4q\") pod \"certified-operators-n6ghn\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.714968 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-utilities\") pod \"certified-operators-n6ghn\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.715448 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-catalog-content\") pod \"certified-operators-n6ghn\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.738033 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mr4q\" (UniqueName: \"kubernetes.io/projected/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-kube-api-access-4mr4q\") pod \"certified-operators-n6ghn\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:08 crc kubenswrapper[4954]: I1209 17:33:08.838185 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:09 crc kubenswrapper[4954]: I1209 17:33:09.345497 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n6ghn"] Dec 09 17:33:09 crc kubenswrapper[4954]: W1209 17:33:09.351762 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fd70caa_6c26_4b74_8c9e_ccda6520bc56.slice/crio-ad44afcc294e775f26762698db16916c1decc03362e242882248e1f4426a55e7 WatchSource:0}: Error finding container ad44afcc294e775f26762698db16916c1decc03362e242882248e1f4426a55e7: Status 404 returned error can't find the container with id ad44afcc294e775f26762698db16916c1decc03362e242882248e1f4426a55e7 Dec 09 17:33:09 crc kubenswrapper[4954]: I1209 17:33:09.551353 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6ghn" event={"ID":"7fd70caa-6c26-4b74-8c9e-ccda6520bc56","Type":"ContainerStarted","Data":"ad44afcc294e775f26762698db16916c1decc03362e242882248e1f4426a55e7"} Dec 09 17:33:10 crc kubenswrapper[4954]: E1209 17:33:10.123655 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:33:10 crc kubenswrapper[4954]: I1209 17:33:10.564529 4954 generic.go:334] "Generic (PLEG): container finished" podID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerID="b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf" exitCode=0 Dec 09 17:33:10 crc kubenswrapper[4954]: I1209 17:33:10.564647 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6ghn" event={"ID":"7fd70caa-6c26-4b74-8c9e-ccda6520bc56","Type":"ContainerDied","Data":"b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf"} Dec 09 17:33:11 crc kubenswrapper[4954]: I1209 17:33:11.445746 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:11 crc kubenswrapper[4954]: I1209 17:33:11.446153 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:11 crc kubenswrapper[4954]: I1209 17:33:11.498613 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:11 crc kubenswrapper[4954]: I1209 17:33:11.581771 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6ghn" event={"ID":"7fd70caa-6c26-4b74-8c9e-ccda6520bc56","Type":"ContainerStarted","Data":"96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d"} Dec 09 17:33:11 crc kubenswrapper[4954]: I1209 17:33:11.631813 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:12 crc kubenswrapper[4954]: I1209 17:33:12.595655 4954 generic.go:334] "Generic (PLEG): container finished" podID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerID="96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d" exitCode=0 Dec 09 17:33:12 crc kubenswrapper[4954]: I1209 17:33:12.595716 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6ghn" event={"ID":"7fd70caa-6c26-4b74-8c9e-ccda6520bc56","Type":"ContainerDied","Data":"96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d"} Dec 09 17:33:13 crc kubenswrapper[4954]: I1209 17:33:13.611728 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6ghn" event={"ID":"7fd70caa-6c26-4b74-8c9e-ccda6520bc56","Type":"ContainerStarted","Data":"9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7"} Dec 09 17:33:13 crc kubenswrapper[4954]: I1209 17:33:13.681893 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n6ghn" podStartSLOduration=3.223362343 podStartE2EDuration="5.681869785s" podCreationTimestamp="2025-12-09 17:33:08 +0000 UTC" firstStartedPulling="2025-12-09 17:33:10.566491143 +0000 UTC m=+2186.954664963" lastFinishedPulling="2025-12-09 17:33:13.024998585 +0000 UTC m=+2189.413172405" observedRunningTime="2025-12-09 17:33:13.638777886 +0000 UTC m=+2190.026951696" watchObservedRunningTime="2025-12-09 17:33:13.681869785 +0000 UTC m=+2190.070043595" Dec 09 17:33:13 crc kubenswrapper[4954]: I1209 17:33:13.691293 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6drgz"] Dec 09 17:33:13 crc kubenswrapper[4954]: I1209 17:33:13.691570 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6drgz" podUID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerName="registry-server" containerID="cri-o://900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5" gracePeriod=2 Dec 09 17:33:13 crc kubenswrapper[4954]: I1209 17:33:13.756656 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:33:13 crc kubenswrapper[4954]: I1209 17:33:13.756721 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:33:13 crc kubenswrapper[4954]: I1209 17:33:13.756765 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:33:13 crc kubenswrapper[4954]: I1209 17:33:13.757820 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f2aaeb263fa3699e3f6e9d98e811f8ee1865140f8b50100f2a0f8b5ce6f4313"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:33:13 crc kubenswrapper[4954]: I1209 17:33:13.757880 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://7f2aaeb263fa3699e3f6e9d98e811f8ee1865140f8b50100f2a0f8b5ce6f4313" gracePeriod=600 Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.315490 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.448728 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcf8f\" (UniqueName: \"kubernetes.io/projected/d214bb4b-5cf1-4d40-9770-01d7bf789052-kube-api-access-bcf8f\") pod \"d214bb4b-5cf1-4d40-9770-01d7bf789052\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.448908 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-utilities\") pod \"d214bb4b-5cf1-4d40-9770-01d7bf789052\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.449250 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-catalog-content\") pod \"d214bb4b-5cf1-4d40-9770-01d7bf789052\" (UID: \"d214bb4b-5cf1-4d40-9770-01d7bf789052\") " Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.449968 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-utilities" (OuterVolumeSpecName: "utilities") pod "d214bb4b-5cf1-4d40-9770-01d7bf789052" (UID: "d214bb4b-5cf1-4d40-9770-01d7bf789052"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.460360 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d214bb4b-5cf1-4d40-9770-01d7bf789052-kube-api-access-bcf8f" (OuterVolumeSpecName: "kube-api-access-bcf8f") pod "d214bb4b-5cf1-4d40-9770-01d7bf789052" (UID: "d214bb4b-5cf1-4d40-9770-01d7bf789052"). InnerVolumeSpecName "kube-api-access-bcf8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.471542 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d214bb4b-5cf1-4d40-9770-01d7bf789052" (UID: "d214bb4b-5cf1-4d40-9770-01d7bf789052"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.558979 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcf8f\" (UniqueName: \"kubernetes.io/projected/d214bb4b-5cf1-4d40-9770-01d7bf789052-kube-api-access-bcf8f\") on node \"crc\" DevicePath \"\"" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.559260 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.559321 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d214bb4b-5cf1-4d40-9770-01d7bf789052-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.639700 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="7f2aaeb263fa3699e3f6e9d98e811f8ee1865140f8b50100f2a0f8b5ce6f4313" exitCode=0 Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.639795 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"7f2aaeb263fa3699e3f6e9d98e811f8ee1865140f8b50100f2a0f8b5ce6f4313"} Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.641792 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b"} Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.641973 4954 scope.go:117] "RemoveContainer" containerID="5fb3683e4cea5fabada45cc368fd94cf0aed9224350eb0a26fd4854f03fe1b0d" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.650691 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6drgz" event={"ID":"d214bb4b-5cf1-4d40-9770-01d7bf789052","Type":"ContainerDied","Data":"900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5"} Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.650760 4954 generic.go:334] "Generic (PLEG): container finished" podID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerID="900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5" exitCode=0 Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.650901 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6drgz" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.651072 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6drgz" event={"ID":"d214bb4b-5cf1-4d40-9770-01d7bf789052","Type":"ContainerDied","Data":"cbb25fd76f8ffd7373842447a4d2b5681f09b41c32991adc6a4020960117312f"} Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.686662 4954 scope.go:117] "RemoveContainer" containerID="900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.709725 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6drgz"] Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.717556 4954 scope.go:117] "RemoveContainer" containerID="6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.734564 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6drgz"] Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.743186 4954 scope.go:117] "RemoveContainer" containerID="f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.799996 4954 scope.go:117] "RemoveContainer" containerID="900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5" Dec 09 17:33:14 crc kubenswrapper[4954]: E1209 17:33:14.800728 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5\": container with ID starting with 900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5 not found: ID does not exist" containerID="900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.800782 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5"} err="failed to get container status \"900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5\": rpc error: code = NotFound desc = could not find container \"900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5\": container with ID starting with 900c823b42c138e7492a57229cc728969da7aa4ac6e28916d4f01431bbee07c5 not found: ID does not exist" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.800818 4954 scope.go:117] "RemoveContainer" containerID="6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3" Dec 09 17:33:14 crc kubenswrapper[4954]: E1209 17:33:14.802897 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3\": container with ID starting with 6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3 not found: ID does not exist" containerID="6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.802933 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3"} err="failed to get container status \"6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3\": rpc error: code = NotFound desc = could not find container \"6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3\": container with ID starting with 6df3a4b11ac7bdff2aa3b2a4e02c82640dd5f7e13179984748730fc190b045c3 not found: ID does not exist" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.802956 4954 scope.go:117] "RemoveContainer" containerID="f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c" Dec 09 17:33:14 crc kubenswrapper[4954]: E1209 17:33:14.803435 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c\": container with ID starting with f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c not found: ID does not exist" containerID="f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c" Dec 09 17:33:14 crc kubenswrapper[4954]: I1209 17:33:14.803505 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c"} err="failed to get container status \"f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c\": rpc error: code = NotFound desc = could not find container \"f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c\": container with ID starting with f22ca68bb28bffe22a20deefa5914339d8b8a4d349cce2a7b9160d64e876536c not found: ID does not exist" Dec 09 17:33:15 crc kubenswrapper[4954]: E1209 17:33:15.243288 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:33:15 crc kubenswrapper[4954]: E1209 17:33:15.243369 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:33:15 crc kubenswrapper[4954]: E1209 17:33:15.243550 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:33:15 crc kubenswrapper[4954]: E1209 17:33:15.244909 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:33:16 crc kubenswrapper[4954]: I1209 17:33:16.133813 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d214bb4b-5cf1-4d40-9770-01d7bf789052" path="/var/lib/kubelet/pods/d214bb4b-5cf1-4d40-9770-01d7bf789052/volumes" Dec 09 17:33:18 crc kubenswrapper[4954]: I1209 17:33:18.839193 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:18 crc kubenswrapper[4954]: I1209 17:33:18.839878 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:18 crc kubenswrapper[4954]: I1209 17:33:18.892294 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:19 crc kubenswrapper[4954]: I1209 17:33:19.767496 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:19 crc kubenswrapper[4954]: I1209 17:33:19.886970 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6ghn"] Dec 09 17:33:21 crc kubenswrapper[4954]: I1209 17:33:21.731898 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n6ghn" podUID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerName="registry-server" containerID="cri-o://9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7" gracePeriod=2 Dec 09 17:33:22 crc kubenswrapper[4954]: E1209 17:33:22.248621 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:33:22 crc kubenswrapper[4954]: E1209 17:33:22.249237 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:33:22 crc kubenswrapper[4954]: E1209 17:33:22.249443 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:33:22 crc kubenswrapper[4954]: E1209 17:33:22.250570 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.253368 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.342122 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-catalog-content\") pod \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.342234 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-utilities\") pod \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.342388 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mr4q\" (UniqueName: \"kubernetes.io/projected/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-kube-api-access-4mr4q\") pod \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\" (UID: \"7fd70caa-6c26-4b74-8c9e-ccda6520bc56\") " Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.343360 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-utilities" (OuterVolumeSpecName: "utilities") pod "7fd70caa-6c26-4b74-8c9e-ccda6520bc56" (UID: "7fd70caa-6c26-4b74-8c9e-ccda6520bc56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.356629 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-kube-api-access-4mr4q" (OuterVolumeSpecName: "kube-api-access-4mr4q") pod "7fd70caa-6c26-4b74-8c9e-ccda6520bc56" (UID: "7fd70caa-6c26-4b74-8c9e-ccda6520bc56"). InnerVolumeSpecName "kube-api-access-4mr4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.395883 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fd70caa-6c26-4b74-8c9e-ccda6520bc56" (UID: "7fd70caa-6c26-4b74-8c9e-ccda6520bc56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.445043 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.445082 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.445096 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mr4q\" (UniqueName: \"kubernetes.io/projected/7fd70caa-6c26-4b74-8c9e-ccda6520bc56-kube-api-access-4mr4q\") on node \"crc\" DevicePath \"\"" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.745072 4954 generic.go:334] "Generic (PLEG): container finished" podID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerID="9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7" exitCode=0 Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.745123 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6ghn" event={"ID":"7fd70caa-6c26-4b74-8c9e-ccda6520bc56","Type":"ContainerDied","Data":"9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7"} Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.745425 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n6ghn" event={"ID":"7fd70caa-6c26-4b74-8c9e-ccda6520bc56","Type":"ContainerDied","Data":"ad44afcc294e775f26762698db16916c1decc03362e242882248e1f4426a55e7"} Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.745448 4954 scope.go:117] "RemoveContainer" containerID="9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.745202 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n6ghn" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.769834 4954 scope.go:117] "RemoveContainer" containerID="96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.794350 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n6ghn"] Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.806325 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n6ghn"] Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.820520 4954 scope.go:117] "RemoveContainer" containerID="b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.878339 4954 scope.go:117] "RemoveContainer" containerID="9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7" Dec 09 17:33:22 crc kubenswrapper[4954]: E1209 17:33:22.879628 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7\": container with ID starting with 9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7 not found: ID does not exist" containerID="9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.879673 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7"} err="failed to get container status \"9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7\": rpc error: code = NotFound desc = could not find container \"9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7\": container with ID starting with 9a9981e99a2c33c389166730a11e6fe7d983b03ec4aba4c00dc1991adfac9ba7 not found: ID does not exist" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.879708 4954 scope.go:117] "RemoveContainer" containerID="96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d" Dec 09 17:33:22 crc kubenswrapper[4954]: E1209 17:33:22.880240 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d\": container with ID starting with 96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d not found: ID does not exist" containerID="96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.880267 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d"} err="failed to get container status \"96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d\": rpc error: code = NotFound desc = could not find container \"96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d\": container with ID starting with 96236fd132c4c218ffa36e779099bca7c68e3f8be4cf2b9a1c62d0831280ea6d not found: ID does not exist" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.880284 4954 scope.go:117] "RemoveContainer" containerID="b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf" Dec 09 17:33:22 crc kubenswrapper[4954]: E1209 17:33:22.880898 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf\": container with ID starting with b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf not found: ID does not exist" containerID="b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf" Dec 09 17:33:22 crc kubenswrapper[4954]: I1209 17:33:22.880925 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf"} err="failed to get container status \"b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf\": rpc error: code = NotFound desc = could not find container \"b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf\": container with ID starting with b6e6661fe3ca36f50aef756838316b3ede11100af755f5be2efd6796fc6b7abf not found: ID does not exist" Dec 09 17:33:24 crc kubenswrapper[4954]: I1209 17:33:24.132502 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" path="/var/lib/kubelet/pods/7fd70caa-6c26-4b74-8c9e-ccda6520bc56/volumes" Dec 09 17:33:26 crc kubenswrapper[4954]: E1209 17:33:26.123353 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:33:33 crc kubenswrapper[4954]: E1209 17:33:33.124524 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:33:37 crc kubenswrapper[4954]: E1209 17:33:37.123100 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:33:48 crc kubenswrapper[4954]: E1209 17:33:48.125623 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:33:49 crc kubenswrapper[4954]: E1209 17:33:49.123280 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:34:02 crc kubenswrapper[4954]: E1209 17:34:02.123362 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:34:02 crc kubenswrapper[4954]: E1209 17:34:02.123462 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:34:02 crc kubenswrapper[4954]: I1209 17:34:02.250569 4954 scope.go:117] "RemoveContainer" containerID="75b288f79920a57cf39dd57ed532721dab0f87ebf88519013a2061e8f811e374" Dec 09 17:34:02 crc kubenswrapper[4954]: I1209 17:34:02.283938 4954 scope.go:117] "RemoveContainer" containerID="ce673116279a476e69c8b4fbe524f61413724f7e556919489ae5879984bf82d7" Dec 09 17:34:14 crc kubenswrapper[4954]: E1209 17:34:14.133496 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:34:16 crc kubenswrapper[4954]: E1209 17:34:16.125107 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:34:29 crc kubenswrapper[4954]: E1209 17:34:29.122853 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:34:29 crc kubenswrapper[4954]: E1209 17:34:29.122909 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:34:40 crc kubenswrapper[4954]: E1209 17:34:40.237658 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:34:40 crc kubenswrapper[4954]: E1209 17:34:40.238239 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:34:40 crc kubenswrapper[4954]: E1209 17:34:40.238397 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:34:40 crc kubenswrapper[4954]: E1209 17:34:40.239623 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:34:43 crc kubenswrapper[4954]: E1209 17:34:43.219755 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:34:43 crc kubenswrapper[4954]: E1209 17:34:43.220347 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:34:43 crc kubenswrapper[4954]: E1209 17:34:43.220524 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:34:43 crc kubenswrapper[4954]: E1209 17:34:43.221749 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:34:54 crc kubenswrapper[4954]: E1209 17:34:54.137873 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:34:54 crc kubenswrapper[4954]: E1209 17:34:54.138749 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:35:05 crc kubenswrapper[4954]: E1209 17:35:05.123782 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:35:08 crc kubenswrapper[4954]: E1209 17:35:08.122551 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:35:18 crc kubenswrapper[4954]: E1209 17:35:18.123841 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:35:19 crc kubenswrapper[4954]: E1209 17:35:19.122791 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:35:31 crc kubenswrapper[4954]: E1209 17:35:31.123262 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:35:33 crc kubenswrapper[4954]: E1209 17:35:33.122706 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:35:42 crc kubenswrapper[4954]: E1209 17:35:42.124379 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:35:43 crc kubenswrapper[4954]: I1209 17:35:43.756528 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:35:43 crc kubenswrapper[4954]: I1209 17:35:43.756951 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:35:44 crc kubenswrapper[4954]: E1209 17:35:44.132924 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:35:53 crc kubenswrapper[4954]: E1209 17:35:53.123774 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:35:58 crc kubenswrapper[4954]: E1209 17:35:58.123570 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:36:07 crc kubenswrapper[4954]: E1209 17:36:07.123829 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:36:10 crc kubenswrapper[4954]: E1209 17:36:10.123299 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:36:13 crc kubenswrapper[4954]: I1209 17:36:13.755715 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:36:13 crc kubenswrapper[4954]: I1209 17:36:13.756014 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:36:15 crc kubenswrapper[4954]: I1209 17:36:15.080168 4954 generic.go:334] "Generic (PLEG): container finished" podID="43bffa1b-06d5-470d-b410-82f6acf14286" containerID="0e4a7881f2c239a6fa3573ee2ef18170972f9b96f504ed856cf328a6cdf2d68d" exitCode=0 Dec 09 17:36:15 crc kubenswrapper[4954]: I1209 17:36:15.080237 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" event={"ID":"43bffa1b-06d5-470d-b410-82f6acf14286","Type":"ContainerDied","Data":"0e4a7881f2c239a6fa3573ee2ef18170972f9b96f504ed856cf328a6cdf2d68d"} Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.532625 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.657153 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-ssh-key\") pod \"43bffa1b-06d5-470d-b410-82f6acf14286\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.657554 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-bootstrap-combined-ca-bundle\") pod \"43bffa1b-06d5-470d-b410-82f6acf14286\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.657614 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-inventory\") pod \"43bffa1b-06d5-470d-b410-82f6acf14286\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.657710 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p5lt\" (UniqueName: \"kubernetes.io/projected/43bffa1b-06d5-470d-b410-82f6acf14286-kube-api-access-2p5lt\") pod \"43bffa1b-06d5-470d-b410-82f6acf14286\" (UID: \"43bffa1b-06d5-470d-b410-82f6acf14286\") " Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.663305 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43bffa1b-06d5-470d-b410-82f6acf14286-kube-api-access-2p5lt" (OuterVolumeSpecName: "kube-api-access-2p5lt") pod "43bffa1b-06d5-470d-b410-82f6acf14286" (UID: "43bffa1b-06d5-470d-b410-82f6acf14286"). InnerVolumeSpecName "kube-api-access-2p5lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.663773 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "43bffa1b-06d5-470d-b410-82f6acf14286" (UID: "43bffa1b-06d5-470d-b410-82f6acf14286"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.690314 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43bffa1b-06d5-470d-b410-82f6acf14286" (UID: "43bffa1b-06d5-470d-b410-82f6acf14286"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.690834 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-inventory" (OuterVolumeSpecName: "inventory") pod "43bffa1b-06d5-470d-b410-82f6acf14286" (UID: "43bffa1b-06d5-470d-b410-82f6acf14286"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.760329 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.760652 4954 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.760729 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43bffa1b-06d5-470d-b410-82f6acf14286-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:36:16 crc kubenswrapper[4954]: I1209 17:36:16.760822 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p5lt\" (UniqueName: \"kubernetes.io/projected/43bffa1b-06d5-470d-b410-82f6acf14286-kube-api-access-2p5lt\") on node \"crc\" DevicePath \"\"" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.104790 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" event={"ID":"43bffa1b-06d5-470d-b410-82f6acf14286","Type":"ContainerDied","Data":"155c4611b6d7c9e35ec6f179430532289dff8153b16852a92da0706129ab61bc"} Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.105137 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="155c4611b6d7c9e35ec6f179430532289dff8153b16852a92da0706129ab61bc" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.104921 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.205657 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv"] Dec 09 17:36:17 crc kubenswrapper[4954]: E1209 17:36:17.206319 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerName="registry-server" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.206354 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerName="registry-server" Dec 09 17:36:17 crc kubenswrapper[4954]: E1209 17:36:17.206382 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerName="extract-utilities" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.206389 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerName="extract-utilities" Dec 09 17:36:17 crc kubenswrapper[4954]: E1209 17:36:17.206408 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerName="extract-content" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.206417 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerName="extract-content" Dec 09 17:36:17 crc kubenswrapper[4954]: E1209 17:36:17.206434 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerName="extract-utilities" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.206441 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerName="extract-utilities" Dec 09 17:36:17 crc kubenswrapper[4954]: E1209 17:36:17.206454 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerName="extract-content" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.206460 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerName="extract-content" Dec 09 17:36:17 crc kubenswrapper[4954]: E1209 17:36:17.206482 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43bffa1b-06d5-470d-b410-82f6acf14286" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.206489 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="43bffa1b-06d5-470d-b410-82f6acf14286" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 17:36:17 crc kubenswrapper[4954]: E1209 17:36:17.206503 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerName="registry-server" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.206509 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerName="registry-server" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.206752 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="43bffa1b-06d5-470d-b410-82f6acf14286" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.206780 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d214bb4b-5cf1-4d40-9770-01d7bf789052" containerName="registry-server" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.206794 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd70caa-6c26-4b74-8c9e-ccda6520bc56" containerName="registry-server" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.207801 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.211285 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.212174 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.214311 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.214364 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.236685 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv"] Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.275325 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.275503 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.275794 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drc78\" (UniqueName: \"kubernetes.io/projected/dbd5865d-2c5b-415f-a824-d0acfefad34a-kube-api-access-drc78\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.377891 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.378107 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drc78\" (UniqueName: \"kubernetes.io/projected/dbd5865d-2c5b-415f-a824-d0acfefad34a-kube-api-access-drc78\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.378131 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.382565 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.394733 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.401057 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drc78\" (UniqueName: \"kubernetes.io/projected/dbd5865d-2c5b-415f-a824-d0acfefad34a-kube-api-access-drc78\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:17 crc kubenswrapper[4954]: I1209 17:36:17.529108 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:36:18 crc kubenswrapper[4954]: I1209 17:36:18.095843 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv"] Dec 09 17:36:18 crc kubenswrapper[4954]: I1209 17:36:18.108633 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 17:36:18 crc kubenswrapper[4954]: I1209 17:36:18.118030 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" event={"ID":"dbd5865d-2c5b-415f-a824-d0acfefad34a","Type":"ContainerStarted","Data":"2f2e3467423f801bca2955d92ddaf8f3fbf913adfb00ba5f2761c67b99795c64"} Dec 09 17:36:19 crc kubenswrapper[4954]: I1209 17:36:19.135087 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" event={"ID":"dbd5865d-2c5b-415f-a824-d0acfefad34a","Type":"ContainerStarted","Data":"0920436aaf8dacc9d4fbd619a06cfba1db081876f114eaa0ae3469aab4710ad9"} Dec 09 17:36:19 crc kubenswrapper[4954]: I1209 17:36:19.167165 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" podStartSLOduration=1.6446902780000001 podStartE2EDuration="2.167131191s" podCreationTimestamp="2025-12-09 17:36:17 +0000 UTC" firstStartedPulling="2025-12-09 17:36:18.108208886 +0000 UTC m=+2374.496382706" lastFinishedPulling="2025-12-09 17:36:18.630649789 +0000 UTC m=+2375.018823619" observedRunningTime="2025-12-09 17:36:19.155087234 +0000 UTC m=+2375.543261054" watchObservedRunningTime="2025-12-09 17:36:19.167131191 +0000 UTC m=+2375.555305011" Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.060522 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0126-account-create-update-z7h68"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.076359 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-01f1-account-create-update-zd7hl"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.094770 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-c4jlk"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.107270 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-8grd9"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.119103 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0126-account-create-update-z7h68"] Dec 09 17:36:21 crc kubenswrapper[4954]: E1209 17:36:21.123279 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.129557 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-5w9lz"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.139840 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-c8s82"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.150328 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a366-account-create-update-zkvjq"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.160253 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-8grd9"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.171278 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-cfdd-account-create-update-tqjv4"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.182848 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-01f1-account-create-update-zd7hl"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.194967 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a366-account-create-update-zkvjq"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.206063 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-5w9lz"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.216428 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-c4jlk"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.227015 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-c8s82"] Dec 09 17:36:21 crc kubenswrapper[4954]: I1209 17:36:21.236455 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-cfdd-account-create-update-tqjv4"] Dec 09 17:36:22 crc kubenswrapper[4954]: I1209 17:36:22.135567 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13cbee2b-45bb-4d9e-8427-bb2574b8d284" path="/var/lib/kubelet/pods/13cbee2b-45bb-4d9e-8427-bb2574b8d284/volumes" Dec 09 17:36:22 crc kubenswrapper[4954]: I1209 17:36:22.137360 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c2c10db-d15d-4fc4-b430-80b7cfb2c401" path="/var/lib/kubelet/pods/3c2c10db-d15d-4fc4-b430-80b7cfb2c401/volumes" Dec 09 17:36:22 crc kubenswrapper[4954]: I1209 17:36:22.138750 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56f3777f-f5d6-4a05-855e-8a28b67a07eb" path="/var/lib/kubelet/pods/56f3777f-f5d6-4a05-855e-8a28b67a07eb/volumes" Dec 09 17:36:22 crc kubenswrapper[4954]: I1209 17:36:22.139735 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="629b86ef-b3c8-445c-aab0-c59fb2359f57" path="/var/lib/kubelet/pods/629b86ef-b3c8-445c-aab0-c59fb2359f57/volumes" Dec 09 17:36:22 crc kubenswrapper[4954]: I1209 17:36:22.141664 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f22851f-5367-4fc7-92b4-6114a4d0449a" path="/var/lib/kubelet/pods/6f22851f-5367-4fc7-92b4-6114a4d0449a/volumes" Dec 09 17:36:22 crc kubenswrapper[4954]: I1209 17:36:22.142735 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad0e11b4-6dcf-437b-8798-1acdacc534ed" path="/var/lib/kubelet/pods/ad0e11b4-6dcf-437b-8798-1acdacc534ed/volumes" Dec 09 17:36:22 crc kubenswrapper[4954]: I1209 17:36:22.144286 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6438189-ff16-4038-ba8d-3ee0972e45a3" path="/var/lib/kubelet/pods/e6438189-ff16-4038-ba8d-3ee0972e45a3/volumes" Dec 09 17:36:22 crc kubenswrapper[4954]: I1209 17:36:22.145909 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4f98d10-48e9-4d4f-abde-aef5db23a39e" path="/var/lib/kubelet/pods/f4f98d10-48e9-4d4f-abde-aef5db23a39e/volumes" Dec 09 17:36:24 crc kubenswrapper[4954]: E1209 17:36:24.131194 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:36:27 crc kubenswrapper[4954]: I1209 17:36:27.031785 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn"] Dec 09 17:36:27 crc kubenswrapper[4954]: I1209 17:36:27.058072 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-gzgnn"] Dec 09 17:36:28 crc kubenswrapper[4954]: I1209 17:36:28.039056 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-2e7f-account-create-update-hh2nj"] Dec 09 17:36:28 crc kubenswrapper[4954]: I1209 17:36:28.056537 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-2e7f-account-create-update-hh2nj"] Dec 09 17:36:28 crc kubenswrapper[4954]: I1209 17:36:28.133754 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16ad1f90-16e6-4bf8-bdee-509182d1e152" path="/var/lib/kubelet/pods/16ad1f90-16e6-4bf8-bdee-509182d1e152/volumes" Dec 09 17:36:28 crc kubenswrapper[4954]: I1209 17:36:28.134421 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3df17d5e-f6cc-46af-b015-8640baaa8d45" path="/var/lib/kubelet/pods/3df17d5e-f6cc-46af-b015-8640baaa8d45/volumes" Dec 09 17:36:35 crc kubenswrapper[4954]: E1209 17:36:35.123467 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:36:36 crc kubenswrapper[4954]: E1209 17:36:36.122726 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:36:43 crc kubenswrapper[4954]: I1209 17:36:43.755438 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:36:43 crc kubenswrapper[4954]: I1209 17:36:43.756055 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:36:43 crc kubenswrapper[4954]: I1209 17:36:43.756109 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:36:43 crc kubenswrapper[4954]: I1209 17:36:43.757195 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:36:43 crc kubenswrapper[4954]: I1209 17:36:43.757256 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" gracePeriod=600 Dec 09 17:36:43 crc kubenswrapper[4954]: E1209 17:36:43.881845 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:36:44 crc kubenswrapper[4954]: I1209 17:36:44.436448 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" exitCode=0 Dec 09 17:36:44 crc kubenswrapper[4954]: I1209 17:36:44.436517 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b"} Dec 09 17:36:44 crc kubenswrapper[4954]: I1209 17:36:44.436580 4954 scope.go:117] "RemoveContainer" containerID="7f2aaeb263fa3699e3f6e9d98e811f8ee1865140f8b50100f2a0f8b5ce6f4313" Dec 09 17:36:44 crc kubenswrapper[4954]: I1209 17:36:44.437381 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:36:44 crc kubenswrapper[4954]: E1209 17:36:44.437852 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:36:49 crc kubenswrapper[4954]: E1209 17:36:49.123859 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:36:50 crc kubenswrapper[4954]: E1209 17:36:50.125374 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.065066 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-8m66z"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.082341 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-tnzh6"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.094625 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-5d75-account-create-update-shqjl"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.113171 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-mkljg"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.124057 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b62d-account-create-update-2jx8d"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.133836 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-51e0-account-create-update-krjpx"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.143775 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-tnzh6"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.158668 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-4af7-account-create-update-pmqz4"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.169281 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-8m66z"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.178447 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-5d75-account-create-update-shqjl"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.188055 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-l48m5"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.197093 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-mkljg"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.207887 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-51e0-account-create-update-krjpx"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.218493 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b62d-account-create-update-2jx8d"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.228081 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-l48m5"] Dec 09 17:36:53 crc kubenswrapper[4954]: I1209 17:36:53.300073 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-4af7-account-create-update-pmqz4"] Dec 09 17:36:54 crc kubenswrapper[4954]: I1209 17:36:54.132866 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e02fe72-06d2-43b5-add1-f6577e3bcc6c" path="/var/lib/kubelet/pods/1e02fe72-06d2-43b5-add1-f6577e3bcc6c/volumes" Dec 09 17:36:54 crc kubenswrapper[4954]: I1209 17:36:54.133649 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36f77ede-a607-4d03-b939-2e8908e27def" path="/var/lib/kubelet/pods/36f77ede-a607-4d03-b939-2e8908e27def/volumes" Dec 09 17:36:54 crc kubenswrapper[4954]: I1209 17:36:54.134356 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74" path="/var/lib/kubelet/pods/6ec8214b-cbe8-4ddf-ac1e-b3e8d5d1ac74/volumes" Dec 09 17:36:54 crc kubenswrapper[4954]: I1209 17:36:54.135068 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75288628-3722-4b4a-a76c-bd448cd5996c" path="/var/lib/kubelet/pods/75288628-3722-4b4a-a76c-bd448cd5996c/volumes" Dec 09 17:36:54 crc kubenswrapper[4954]: I1209 17:36:54.136212 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c6703cc-4185-4dac-ab05-40f953f09e2e" path="/var/lib/kubelet/pods/8c6703cc-4185-4dac-ab05-40f953f09e2e/volumes" Dec 09 17:36:54 crc kubenswrapper[4954]: I1209 17:36:54.136886 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfcb15dd-bfcb-4519-917e-b41df5be28a9" path="/var/lib/kubelet/pods/bfcb15dd-bfcb-4519-917e-b41df5be28a9/volumes" Dec 09 17:36:54 crc kubenswrapper[4954]: I1209 17:36:54.137467 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5b0ee6c-24bf-4b3f-875c-065c7f51862b" path="/var/lib/kubelet/pods/d5b0ee6c-24bf-4b3f-875c-065c7f51862b/volumes" Dec 09 17:36:54 crc kubenswrapper[4954]: I1209 17:36:54.138521 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f32a6b81-b028-4e1d-b32b-2d8271d13875" path="/var/lib/kubelet/pods/f32a6b81-b028-4e1d-b32b-2d8271d13875/volumes" Dec 09 17:36:57 crc kubenswrapper[4954]: I1209 17:36:57.059775 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-g4kjd"] Dec 09 17:36:57 crc kubenswrapper[4954]: I1209 17:36:57.070434 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-g4kjd"] Dec 09 17:36:57 crc kubenswrapper[4954]: I1209 17:36:57.120922 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:36:57 crc kubenswrapper[4954]: E1209 17:36:57.121255 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:36:58 crc kubenswrapper[4954]: I1209 17:36:58.133436 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8640fe28-2072-49f7-8c8d-35a2ea5432e0" path="/var/lib/kubelet/pods/8640fe28-2072-49f7-8c8d-35a2ea5432e0/volumes" Dec 09 17:36:59 crc kubenswrapper[4954]: I1209 17:36:59.034655 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wq9lg"] Dec 09 17:36:59 crc kubenswrapper[4954]: I1209 17:36:59.046156 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wq9lg"] Dec 09 17:37:00 crc kubenswrapper[4954]: I1209 17:37:00.134133 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b75d1898-5670-4f9e-88d9-69e4a1348c95" path="/var/lib/kubelet/pods/b75d1898-5670-4f9e-88d9-69e4a1348c95/volumes" Dec 09 17:37:02 crc kubenswrapper[4954]: E1209 17:37:02.123194 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:37:02 crc kubenswrapper[4954]: I1209 17:37:02.505501 4954 scope.go:117] "RemoveContainer" containerID="746184e81b7f4df22f167becd076855f55e80b82a88c67b894c381c978b3d0b5" Dec 09 17:37:02 crc kubenswrapper[4954]: I1209 17:37:02.549963 4954 scope.go:117] "RemoveContainer" containerID="1c7e0858f76d7fed7b2b05e3a15cea9ed4f931302180a0ed8bfffc3879ada9d5" Dec 09 17:37:02 crc kubenswrapper[4954]: I1209 17:37:02.626337 4954 scope.go:117] "RemoveContainer" containerID="667162aaff42650f3f21cd35f3d7d9f258df7e8210d1932a36d2c2ecf9ce04bd" Dec 09 17:37:02 crc kubenswrapper[4954]: I1209 17:37:02.694350 4954 scope.go:117] "RemoveContainer" containerID="d9e65b648349a288fb93f22404ff0b315eeca9b04e2cd46e7428d85e95c4132d" Dec 09 17:37:02 crc kubenswrapper[4954]: I1209 17:37:02.739114 4954 scope.go:117] "RemoveContainer" containerID="ce5b946d2591077b1daed5f36d9e99e25a3527cf293ef77ee6dc747fb18ebf7b" Dec 09 17:37:02 crc kubenswrapper[4954]: I1209 17:37:02.763250 4954 scope.go:117] "RemoveContainer" containerID="cb0010c4e2f2f22c9accf9a84fe0a7fbb6c8a452ca2e1ef98edfb077874d7679" Dec 09 17:37:02 crc kubenswrapper[4954]: I1209 17:37:02.821511 4954 scope.go:117] "RemoveContainer" containerID="eab74493a154b3368fc126aa75e9d1753a9220c6bf6879cabc7ff20617476208" Dec 09 17:37:02 crc kubenswrapper[4954]: I1209 17:37:02.911018 4954 scope.go:117] "RemoveContainer" containerID="cca583698e0b90d3649182f307906fd96b34a47fdc2eef0599292f61ac2a559a" Dec 09 17:37:02 crc kubenswrapper[4954]: I1209 17:37:02.951435 4954 scope.go:117] "RemoveContainer" containerID="4be6620af5968a980115be19bd7c8ebe122982fe16e90eb212a05e310652bf55" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.000526 4954 scope.go:117] "RemoveContainer" containerID="c602f8328ed1c607f7745f368d5ab19113a48a22ff9962ea7b65c3f7faf3f627" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.055298 4954 scope.go:117] "RemoveContainer" containerID="46bd1e0d8bdb3d52b824343c9b64b4c76bd9290fcee7eb08a2b31393f1461bac" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.091765 4954 scope.go:117] "RemoveContainer" containerID="c22fbc0e99b17e5023a9950dcfa5eda459b4dbbfa399207e2360dc07acd4753e" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.115265 4954 scope.go:117] "RemoveContainer" containerID="9d55d5d0a551d4efa78f31acb42645047cd58b12a48c7b9fd0bcd9b737fcf22c" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.140031 4954 scope.go:117] "RemoveContainer" containerID="d237c59ecd021445baf479358507e0aef7babd5f0631c0c91db2eda378dfbbdf" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.169564 4954 scope.go:117] "RemoveContainer" containerID="9a0a0d8331b1e11d765c8c42cef5cbf36ba8a024bf35353a76a4ede6422ab865" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.195577 4954 scope.go:117] "RemoveContainer" containerID="b23fc88a6f35a9ec77252dc7911681f4bb0b40fb65b51e924b5b61dc0ffe0f51" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.222732 4954 scope.go:117] "RemoveContainer" containerID="4a158212a46076db0596674f8ee27ab9ddc5463498b0dc80de8a368ee4cb82d2" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.244851 4954 scope.go:117] "RemoveContainer" containerID="8a70d9a3e5ab03784ccf64ebf5ffcb86fa80998a545ca58f6941e2039ba6b627" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.270473 4954 scope.go:117] "RemoveContainer" containerID="bd85f23b3130bdeb6dbfa020ea1e347b46593c4e137ef77a95d4897eba77da80" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.296560 4954 scope.go:117] "RemoveContainer" containerID="550c11f7d963f566535d572ec244a4e3166b6b834be5d43a92cc18acd976a737" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.317388 4954 scope.go:117] "RemoveContainer" containerID="ff078a7a54cf55d551ed7b1cc84fd4dc3a7ee136d9797be4beb31a5c0b4022f6" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.341451 4954 scope.go:117] "RemoveContainer" containerID="77d1027a9b3778aa9e7f775fd239898d1f2e67f789ec6871e65de66c17f05eb0" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.367188 4954 scope.go:117] "RemoveContainer" containerID="fd9533e90f600c7b5adf3df73dca386043e8140bad943067cb92af7829fa878f" Dec 09 17:37:03 crc kubenswrapper[4954]: I1209 17:37:03.390291 4954 scope.go:117] "RemoveContainer" containerID="81dbcbdd15ecbd20c8cecc3dc608ef732a88b95f9097ed37cb3f6be09ef975a6" Dec 09 17:37:04 crc kubenswrapper[4954]: E1209 17:37:04.128973 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:37:08 crc kubenswrapper[4954]: I1209 17:37:08.120795 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:37:08 crc kubenswrapper[4954]: E1209 17:37:08.121812 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:37:16 crc kubenswrapper[4954]: E1209 17:37:16.122991 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:37:16 crc kubenswrapper[4954]: E1209 17:37:16.123726 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:37:20 crc kubenswrapper[4954]: I1209 17:37:20.120996 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:37:20 crc kubenswrapper[4954]: E1209 17:37:20.121819 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:37:27 crc kubenswrapper[4954]: E1209 17:37:27.255027 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:37:27 crc kubenswrapper[4954]: E1209 17:37:27.255382 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:37:27 crc kubenswrapper[4954]: E1209 17:37:27.255585 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:37:27 crc kubenswrapper[4954]: E1209 17:37:27.256842 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:37:29 crc kubenswrapper[4954]: E1209 17:37:29.253477 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:37:29 crc kubenswrapper[4954]: E1209 17:37:29.254055 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:37:29 crc kubenswrapper[4954]: E1209 17:37:29.254190 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:37:29 crc kubenswrapper[4954]: E1209 17:37:29.255392 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:37:30 crc kubenswrapper[4954]: I1209 17:37:30.056633 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-cqb6p"] Dec 09 17:37:30 crc kubenswrapper[4954]: I1209 17:37:30.068647 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-cqb6p"] Dec 09 17:37:30 crc kubenswrapper[4954]: I1209 17:37:30.134926 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4296b9c8-795b-4786-80a8-1b8c8d723451" path="/var/lib/kubelet/pods/4296b9c8-795b-4786-80a8-1b8c8d723451/volumes" Dec 09 17:37:34 crc kubenswrapper[4954]: I1209 17:37:34.960212 4954 generic.go:334] "Generic (PLEG): container finished" podID="dbd5865d-2c5b-415f-a824-d0acfefad34a" containerID="0920436aaf8dacc9d4fbd619a06cfba1db081876f114eaa0ae3469aab4710ad9" exitCode=0 Dec 09 17:37:34 crc kubenswrapper[4954]: I1209 17:37:34.960308 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" event={"ID":"dbd5865d-2c5b-415f-a824-d0acfefad34a","Type":"ContainerDied","Data":"0920436aaf8dacc9d4fbd619a06cfba1db081876f114eaa0ae3469aab4710ad9"} Dec 09 17:37:35 crc kubenswrapper[4954]: I1209 17:37:35.120520 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:37:35 crc kubenswrapper[4954]: E1209 17:37:35.121314 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.548879 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.678609 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-ssh-key\") pod \"dbd5865d-2c5b-415f-a824-d0acfefad34a\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.678758 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-inventory\") pod \"dbd5865d-2c5b-415f-a824-d0acfefad34a\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.678945 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drc78\" (UniqueName: \"kubernetes.io/projected/dbd5865d-2c5b-415f-a824-d0acfefad34a-kube-api-access-drc78\") pod \"dbd5865d-2c5b-415f-a824-d0acfefad34a\" (UID: \"dbd5865d-2c5b-415f-a824-d0acfefad34a\") " Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.685618 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbd5865d-2c5b-415f-a824-d0acfefad34a-kube-api-access-drc78" (OuterVolumeSpecName: "kube-api-access-drc78") pod "dbd5865d-2c5b-415f-a824-d0acfefad34a" (UID: "dbd5865d-2c5b-415f-a824-d0acfefad34a"). InnerVolumeSpecName "kube-api-access-drc78". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.710061 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dbd5865d-2c5b-415f-a824-d0acfefad34a" (UID: "dbd5865d-2c5b-415f-a824-d0acfefad34a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.715742 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-inventory" (OuterVolumeSpecName: "inventory") pod "dbd5865d-2c5b-415f-a824-d0acfefad34a" (UID: "dbd5865d-2c5b-415f-a824-d0acfefad34a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.781956 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drc78\" (UniqueName: \"kubernetes.io/projected/dbd5865d-2c5b-415f-a824-d0acfefad34a-kube-api-access-drc78\") on node \"crc\" DevicePath \"\"" Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.781997 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.782007 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbd5865d-2c5b-415f-a824-d0acfefad34a-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.982804 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" event={"ID":"dbd5865d-2c5b-415f-a824-d0acfefad34a","Type":"ContainerDied","Data":"2f2e3467423f801bca2955d92ddaf8f3fbf913adfb00ba5f2761c67b99795c64"} Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.983096 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f2e3467423f801bca2955d92ddaf8f3fbf913adfb00ba5f2761c67b99795c64" Dec 09 17:37:36 crc kubenswrapper[4954]: I1209 17:37:36.982864 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.089343 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4"] Dec 09 17:37:37 crc kubenswrapper[4954]: E1209 17:37:37.101236 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd5865d-2c5b-415f-a824-d0acfefad34a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.101283 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd5865d-2c5b-415f-a824-d0acfefad34a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.101793 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbd5865d-2c5b-415f-a824-d0acfefad34a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.103978 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4"] Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.104147 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.107177 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.107635 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.107764 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.107795 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.293359 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvrnp\" (UniqueName: \"kubernetes.io/projected/5f717719-b2db-4a33-9607-dc210c19261f-kube-api-access-dvrnp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cplh4\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.294874 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cplh4\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.295412 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cplh4\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.398085 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cplh4\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.398182 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvrnp\" (UniqueName: \"kubernetes.io/projected/5f717719-b2db-4a33-9607-dc210c19261f-kube-api-access-dvrnp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cplh4\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.398300 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cplh4\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.403319 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cplh4\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.405198 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cplh4\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.415235 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvrnp\" (UniqueName: \"kubernetes.io/projected/5f717719-b2db-4a33-9607-dc210c19261f-kube-api-access-dvrnp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-cplh4\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:37 crc kubenswrapper[4954]: I1209 17:37:37.433188 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:38 crc kubenswrapper[4954]: I1209 17:37:38.003979 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4"] Dec 09 17:37:39 crc kubenswrapper[4954]: I1209 17:37:39.003158 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" event={"ID":"5f717719-b2db-4a33-9607-dc210c19261f","Type":"ContainerStarted","Data":"330358c26ce11f1a113640a6b471b59b4c521374a2442880f73e5ac781f6bc2b"} Dec 09 17:37:39 crc kubenswrapper[4954]: I1209 17:37:39.003461 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" event={"ID":"5f717719-b2db-4a33-9607-dc210c19261f","Type":"ContainerStarted","Data":"b4b01be61acfb2e9c61ae6ece08143079da4b42ec60200b7c4999c426e36ac62"} Dec 09 17:37:39 crc kubenswrapper[4954]: I1209 17:37:39.034273 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" podStartSLOduration=1.533946474 podStartE2EDuration="2.034245099s" podCreationTimestamp="2025-12-09 17:37:37 +0000 UTC" firstStartedPulling="2025-12-09 17:37:38.010238362 +0000 UTC m=+2454.398412182" lastFinishedPulling="2025-12-09 17:37:38.510536987 +0000 UTC m=+2454.898710807" observedRunningTime="2025-12-09 17:37:39.022632633 +0000 UTC m=+2455.410806463" watchObservedRunningTime="2025-12-09 17:37:39.034245099 +0000 UTC m=+2455.422419279" Dec 09 17:37:41 crc kubenswrapper[4954]: E1209 17:37:41.123052 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:37:43 crc kubenswrapper[4954]: I1209 17:37:43.030390 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-fsxd4"] Dec 09 17:37:43 crc kubenswrapper[4954]: I1209 17:37:43.040247 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-cb6lc"] Dec 09 17:37:43 crc kubenswrapper[4954]: I1209 17:37:43.051438 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-fsxd4"] Dec 09 17:37:43 crc kubenswrapper[4954]: I1209 17:37:43.062671 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-cb6lc"] Dec 09 17:37:44 crc kubenswrapper[4954]: I1209 17:37:44.048155 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-c686v"] Dec 09 17:37:44 crc kubenswrapper[4954]: I1209 17:37:44.065053 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-c686v"] Dec 09 17:37:44 crc kubenswrapper[4954]: I1209 17:37:44.066989 4954 generic.go:334] "Generic (PLEG): container finished" podID="5f717719-b2db-4a33-9607-dc210c19261f" containerID="330358c26ce11f1a113640a6b471b59b4c521374a2442880f73e5ac781f6bc2b" exitCode=0 Dec 09 17:37:44 crc kubenswrapper[4954]: I1209 17:37:44.067043 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" event={"ID":"5f717719-b2db-4a33-9607-dc210c19261f","Type":"ContainerDied","Data":"330358c26ce11f1a113640a6b471b59b4c521374a2442880f73e5ac781f6bc2b"} Dec 09 17:37:44 crc kubenswrapper[4954]: I1209 17:37:44.143326 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b087316-6f76-4aaa-bd43-d5b01051d4b2" path="/var/lib/kubelet/pods/3b087316-6f76-4aaa-bd43-d5b01051d4b2/volumes" Dec 09 17:37:44 crc kubenswrapper[4954]: E1209 17:37:44.143881 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:37:44 crc kubenswrapper[4954]: I1209 17:37:44.144059 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fd76337-c1f2-43ff-b021-8837dd046dc2" path="/var/lib/kubelet/pods/5fd76337-c1f2-43ff-b021-8837dd046dc2/volumes" Dec 09 17:37:44 crc kubenswrapper[4954]: I1209 17:37:44.144772 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0d58b11-2dce-48e1-9445-1c30c431b731" path="/var/lib/kubelet/pods/d0d58b11-2dce-48e1-9445-1c30c431b731/volumes" Dec 09 17:37:45 crc kubenswrapper[4954]: I1209 17:37:45.710967 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:45 crc kubenswrapper[4954]: I1209 17:37:45.817565 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-inventory\") pod \"5f717719-b2db-4a33-9607-dc210c19261f\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " Dec 09 17:37:45 crc kubenswrapper[4954]: I1209 17:37:45.817628 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-ssh-key\") pod \"5f717719-b2db-4a33-9607-dc210c19261f\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " Dec 09 17:37:45 crc kubenswrapper[4954]: I1209 17:37:45.817686 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvrnp\" (UniqueName: \"kubernetes.io/projected/5f717719-b2db-4a33-9607-dc210c19261f-kube-api-access-dvrnp\") pod \"5f717719-b2db-4a33-9607-dc210c19261f\" (UID: \"5f717719-b2db-4a33-9607-dc210c19261f\") " Dec 09 17:37:45 crc kubenswrapper[4954]: I1209 17:37:45.823754 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f717719-b2db-4a33-9607-dc210c19261f-kube-api-access-dvrnp" (OuterVolumeSpecName: "kube-api-access-dvrnp") pod "5f717719-b2db-4a33-9607-dc210c19261f" (UID: "5f717719-b2db-4a33-9607-dc210c19261f"). InnerVolumeSpecName "kube-api-access-dvrnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:37:45 crc kubenswrapper[4954]: I1209 17:37:45.850742 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-inventory" (OuterVolumeSpecName: "inventory") pod "5f717719-b2db-4a33-9607-dc210c19261f" (UID: "5f717719-b2db-4a33-9607-dc210c19261f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:37:45 crc kubenswrapper[4954]: I1209 17:37:45.856608 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5f717719-b2db-4a33-9607-dc210c19261f" (UID: "5f717719-b2db-4a33-9607-dc210c19261f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:37:45 crc kubenswrapper[4954]: I1209 17:37:45.919869 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:37:45 crc kubenswrapper[4954]: I1209 17:37:45.920329 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f717719-b2db-4a33-9607-dc210c19261f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:37:45 crc kubenswrapper[4954]: I1209 17:37:45.920422 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvrnp\" (UniqueName: \"kubernetes.io/projected/5f717719-b2db-4a33-9607-dc210c19261f-kube-api-access-dvrnp\") on node \"crc\" DevicePath \"\"" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.091447 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" event={"ID":"5f717719-b2db-4a33-9607-dc210c19261f","Type":"ContainerDied","Data":"b4b01be61acfb2e9c61ae6ece08143079da4b42ec60200b7c4999c426e36ac62"} Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.091494 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4b01be61acfb2e9c61ae6ece08143079da4b42ec60200b7c4999c426e36ac62" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.091542 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-cplh4" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.187983 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk"] Dec 09 17:37:46 crc kubenswrapper[4954]: E1209 17:37:46.188606 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f717719-b2db-4a33-9607-dc210c19261f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.188630 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f717719-b2db-4a33-9607-dc210c19261f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.188891 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f717719-b2db-4a33-9607-dc210c19261f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.189935 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.192682 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.192994 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.193278 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.193746 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.198481 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk"] Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.224704 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmn2k\" (UniqueName: \"kubernetes.io/projected/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-kube-api-access-dmn2k\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mg9sk\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.224842 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mg9sk\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.224882 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mg9sk\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.328110 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmn2k\" (UniqueName: \"kubernetes.io/projected/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-kube-api-access-dmn2k\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mg9sk\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.330398 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mg9sk\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.330659 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mg9sk\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.335123 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mg9sk\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.335203 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mg9sk\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.347236 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmn2k\" (UniqueName: \"kubernetes.io/projected/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-kube-api-access-dmn2k\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mg9sk\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:46 crc kubenswrapper[4954]: I1209 17:37:46.541787 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:37:47 crc kubenswrapper[4954]: I1209 17:37:47.076056 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk"] Dec 09 17:37:47 crc kubenswrapper[4954]: I1209 17:37:47.103506 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" event={"ID":"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3","Type":"ContainerStarted","Data":"090e84a9040052804614b5b06e737f20abcd6d5cdf5bbee46949b2f2baa97d22"} Dec 09 17:37:48 crc kubenswrapper[4954]: I1209 17:37:48.118634 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" event={"ID":"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3","Type":"ContainerStarted","Data":"a886a77b170aa50faf1da8eeca5b7585a1c6db528332e2de907951b191c35eda"} Dec 09 17:37:48 crc kubenswrapper[4954]: I1209 17:37:48.120262 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:37:48 crc kubenswrapper[4954]: E1209 17:37:48.120653 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:37:48 crc kubenswrapper[4954]: I1209 17:37:48.148770 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" podStartSLOduration=1.6959787990000001 podStartE2EDuration="2.148710688s" podCreationTimestamp="2025-12-09 17:37:46 +0000 UTC" firstStartedPulling="2025-12-09 17:37:47.079958204 +0000 UTC m=+2463.468132024" lastFinishedPulling="2025-12-09 17:37:47.532690093 +0000 UTC m=+2463.920863913" observedRunningTime="2025-12-09 17:37:48.138978613 +0000 UTC m=+2464.527152463" watchObservedRunningTime="2025-12-09 17:37:48.148710688 +0000 UTC m=+2464.536884508" Dec 09 17:37:53 crc kubenswrapper[4954]: E1209 17:37:53.124191 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:37:56 crc kubenswrapper[4954]: I1209 17:37:56.049409 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-r98b8"] Dec 09 17:37:56 crc kubenswrapper[4954]: I1209 17:37:56.060337 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-r98b8"] Dec 09 17:37:56 crc kubenswrapper[4954]: I1209 17:37:56.131886 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9653f9d1-0297-458e-bfb6-b6801ef96c3c" path="/var/lib/kubelet/pods/9653f9d1-0297-458e-bfb6-b6801ef96c3c/volumes" Dec 09 17:37:57 crc kubenswrapper[4954]: E1209 17:37:57.123267 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:38:00 crc kubenswrapper[4954]: I1209 17:38:00.120618 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:38:00 crc kubenswrapper[4954]: E1209 17:38:00.121153 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:38:03 crc kubenswrapper[4954]: I1209 17:38:03.777840 4954 scope.go:117] "RemoveContainer" containerID="9ce2957549eff70070e4dc18369586f0756e0c26b3c09557904e5f37ab26085c" Dec 09 17:38:03 crc kubenswrapper[4954]: I1209 17:38:03.840330 4954 scope.go:117] "RemoveContainer" containerID="3ed51eac127f0f876dc383fb126166c8f78932579d8696ec733514adecd2268a" Dec 09 17:38:03 crc kubenswrapper[4954]: I1209 17:38:03.877971 4954 scope.go:117] "RemoveContainer" containerID="4104c8ec76bad3c56a6c514d99d167e0bb6322b543aa3dbff81c00a9d841ec15" Dec 09 17:38:03 crc kubenswrapper[4954]: I1209 17:38:03.956375 4954 scope.go:117] "RemoveContainer" containerID="bc9fa86c3f1653c3b63b8f3213022e39348d9149dae183557f2a8bed522b9738" Dec 09 17:38:04 crc kubenswrapper[4954]: I1209 17:38:04.000484 4954 scope.go:117] "RemoveContainer" containerID="8eaf123710b9afd215b706b5e73e57f168af7d51f06ec2b030a11171460b9ffa" Dec 09 17:38:08 crc kubenswrapper[4954]: E1209 17:38:08.123362 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:38:11 crc kubenswrapper[4954]: E1209 17:38:11.123288 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:38:13 crc kubenswrapper[4954]: I1209 17:38:13.121009 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:38:13 crc kubenswrapper[4954]: E1209 17:38:13.122143 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:38:22 crc kubenswrapper[4954]: E1209 17:38:22.123914 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:38:23 crc kubenswrapper[4954]: E1209 17:38:23.124089 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:38:24 crc kubenswrapper[4954]: I1209 17:38:24.127204 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:38:24 crc kubenswrapper[4954]: E1209 17:38:24.127573 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:38:26 crc kubenswrapper[4954]: I1209 17:38:26.505585 4954 generic.go:334] "Generic (PLEG): container finished" podID="511ba2be-4b0f-4d57-8ff8-d2a25f541fd3" containerID="a886a77b170aa50faf1da8eeca5b7585a1c6db528332e2de907951b191c35eda" exitCode=0 Dec 09 17:38:26 crc kubenswrapper[4954]: I1209 17:38:26.505818 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" event={"ID":"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3","Type":"ContainerDied","Data":"a886a77b170aa50faf1da8eeca5b7585a1c6db528332e2de907951b191c35eda"} Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.071215 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.193580 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmn2k\" (UniqueName: \"kubernetes.io/projected/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-kube-api-access-dmn2k\") pod \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.193947 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-inventory\") pod \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.194082 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-ssh-key\") pod \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\" (UID: \"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3\") " Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.200334 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-kube-api-access-dmn2k" (OuterVolumeSpecName: "kube-api-access-dmn2k") pod "511ba2be-4b0f-4d57-8ff8-d2a25f541fd3" (UID: "511ba2be-4b0f-4d57-8ff8-d2a25f541fd3"). InnerVolumeSpecName "kube-api-access-dmn2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.225314 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-inventory" (OuterVolumeSpecName: "inventory") pod "511ba2be-4b0f-4d57-8ff8-d2a25f541fd3" (UID: "511ba2be-4b0f-4d57-8ff8-d2a25f541fd3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.226831 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "511ba2be-4b0f-4d57-8ff8-d2a25f541fd3" (UID: "511ba2be-4b0f-4d57-8ff8-d2a25f541fd3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.297126 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.297452 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.297614 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmn2k\" (UniqueName: \"kubernetes.io/projected/511ba2be-4b0f-4d57-8ff8-d2a25f541fd3-kube-api-access-dmn2k\") on node \"crc\" DevicePath \"\"" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.611866 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" event={"ID":"511ba2be-4b0f-4d57-8ff8-d2a25f541fd3","Type":"ContainerDied","Data":"090e84a9040052804614b5b06e737f20abcd6d5cdf5bbee46949b2f2baa97d22"} Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.611939 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="090e84a9040052804614b5b06e737f20abcd6d5cdf5bbee46949b2f2baa97d22" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.612022 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mg9sk" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.685028 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7"] Dec 09 17:38:28 crc kubenswrapper[4954]: E1209 17:38:28.685576 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="511ba2be-4b0f-4d57-8ff8-d2a25f541fd3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.685676 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="511ba2be-4b0f-4d57-8ff8-d2a25f541fd3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.685953 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="511ba2be-4b0f-4d57-8ff8-d2a25f541fd3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.686834 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.695430 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7"] Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.740115 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.740522 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.740708 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.740830 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.744812 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msg9f\" (UniqueName: \"kubernetes.io/projected/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-kube-api-access-msg9f\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.745177 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.745208 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.847026 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msg9f\" (UniqueName: \"kubernetes.io/projected/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-kube-api-access-msg9f\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.847089 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.847129 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.852406 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.864628 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:28 crc kubenswrapper[4954]: I1209 17:38:28.883402 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msg9f\" (UniqueName: \"kubernetes.io/projected/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-kube-api-access-msg9f\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:29 crc kubenswrapper[4954]: I1209 17:38:29.061125 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:29 crc kubenswrapper[4954]: I1209 17:38:29.663865 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7"] Dec 09 17:38:30 crc kubenswrapper[4954]: I1209 17:38:30.631224 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" event={"ID":"1062cfab-9c0f-4965-a8bf-75ff7f1744d2","Type":"ContainerStarted","Data":"f6e6b99a83a27c49f997570251fdf72d8d31c4789391bce5834a3b60ad9ebac0"} Dec 09 17:38:31 crc kubenswrapper[4954]: I1209 17:38:31.644214 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" event={"ID":"1062cfab-9c0f-4965-a8bf-75ff7f1744d2","Type":"ContainerStarted","Data":"c0ca68ea36893967ab642e3ac99beb8c9202074ff8b3ec567e5e644d90e1e3ee"} Dec 09 17:38:31 crc kubenswrapper[4954]: I1209 17:38:31.671759 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" podStartSLOduration=3.176251176 podStartE2EDuration="3.671737421s" podCreationTimestamp="2025-12-09 17:38:28 +0000 UTC" firstStartedPulling="2025-12-09 17:38:29.672368956 +0000 UTC m=+2506.060542786" lastFinishedPulling="2025-12-09 17:38:30.167855211 +0000 UTC m=+2506.556029031" observedRunningTime="2025-12-09 17:38:31.66313512 +0000 UTC m=+2508.051308940" watchObservedRunningTime="2025-12-09 17:38:31.671737421 +0000 UTC m=+2508.059911241" Dec 09 17:38:34 crc kubenswrapper[4954]: E1209 17:38:34.129482 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:38:34 crc kubenswrapper[4954]: E1209 17:38:34.129914 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:38:34 crc kubenswrapper[4954]: I1209 17:38:34.672898 4954 generic.go:334] "Generic (PLEG): container finished" podID="1062cfab-9c0f-4965-a8bf-75ff7f1744d2" containerID="c0ca68ea36893967ab642e3ac99beb8c9202074ff8b3ec567e5e644d90e1e3ee" exitCode=0 Dec 09 17:38:34 crc kubenswrapper[4954]: I1209 17:38:34.672959 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" event={"ID":"1062cfab-9c0f-4965-a8bf-75ff7f1744d2","Type":"ContainerDied","Data":"c0ca68ea36893967ab642e3ac99beb8c9202074ff8b3ec567e5e644d90e1e3ee"} Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.143972 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.255262 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msg9f\" (UniqueName: \"kubernetes.io/projected/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-kube-api-access-msg9f\") pod \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.255673 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-ssh-key\") pod \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.255873 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-inventory\") pod \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\" (UID: \"1062cfab-9c0f-4965-a8bf-75ff7f1744d2\") " Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.261400 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-kube-api-access-msg9f" (OuterVolumeSpecName: "kube-api-access-msg9f") pod "1062cfab-9c0f-4965-a8bf-75ff7f1744d2" (UID: "1062cfab-9c0f-4965-a8bf-75ff7f1744d2"). InnerVolumeSpecName "kube-api-access-msg9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.284581 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-inventory" (OuterVolumeSpecName: "inventory") pod "1062cfab-9c0f-4965-a8bf-75ff7f1744d2" (UID: "1062cfab-9c0f-4965-a8bf-75ff7f1744d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.285537 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1062cfab-9c0f-4965-a8bf-75ff7f1744d2" (UID: "1062cfab-9c0f-4965-a8bf-75ff7f1744d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.361849 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.361896 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msg9f\" (UniqueName: \"kubernetes.io/projected/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-kube-api-access-msg9f\") on node \"crc\" DevicePath \"\"" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.361913 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1062cfab-9c0f-4965-a8bf-75ff7f1744d2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.696517 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" event={"ID":"1062cfab-9c0f-4965-a8bf-75ff7f1744d2","Type":"ContainerDied","Data":"f6e6b99a83a27c49f997570251fdf72d8d31c4789391bce5834a3b60ad9ebac0"} Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.696551 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.696555 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6e6b99a83a27c49f997570251fdf72d8d31c4789391bce5834a3b60ad9ebac0" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.768715 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8"] Dec 09 17:38:36 crc kubenswrapper[4954]: E1209 17:38:36.769309 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1062cfab-9c0f-4965-a8bf-75ff7f1744d2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.769329 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="1062cfab-9c0f-4965-a8bf-75ff7f1744d2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.769577 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="1062cfab-9c0f-4965-a8bf-75ff7f1744d2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.770566 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.773103 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.773451 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.773585 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.775925 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.782362 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8"] Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.870222 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhvrm\" (UniqueName: \"kubernetes.io/projected/b6a4b196-e34d-4b20-9915-f290a4dac804-kube-api-access-hhvrm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.870549 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.870802 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.973510 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.973587 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.973729 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhvrm\" (UniqueName: \"kubernetes.io/projected/b6a4b196-e34d-4b20-9915-f290a4dac804-kube-api-access-hhvrm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.977104 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.977156 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:36 crc kubenswrapper[4954]: I1209 17:38:36.992716 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhvrm\" (UniqueName: \"kubernetes.io/projected/b6a4b196-e34d-4b20-9915-f290a4dac804-kube-api-access-hhvrm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:37 crc kubenswrapper[4954]: I1209 17:38:37.089018 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:38:37 crc kubenswrapper[4954]: I1209 17:38:37.661877 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8"] Dec 09 17:38:37 crc kubenswrapper[4954]: I1209 17:38:37.707184 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" event={"ID":"b6a4b196-e34d-4b20-9915-f290a4dac804","Type":"ContainerStarted","Data":"0cb35adfae29332c69edf7df0999f2b187d74c1357ecb951230b2320f8d207f6"} Dec 09 17:38:38 crc kubenswrapper[4954]: I1209 17:38:38.120911 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:38:38 crc kubenswrapper[4954]: E1209 17:38:38.121398 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:38:38 crc kubenswrapper[4954]: I1209 17:38:38.719353 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" event={"ID":"b6a4b196-e34d-4b20-9915-f290a4dac804","Type":"ContainerStarted","Data":"32f0c146d7dd24e54a8e0fc654d88db6d1ecbdfce7b2c06834e1b3d8d5114754"} Dec 09 17:38:45 crc kubenswrapper[4954]: E1209 17:38:45.124063 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:38:46 crc kubenswrapper[4954]: I1209 17:38:46.034579 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" podStartSLOduration=9.352209739 podStartE2EDuration="10.03455431s" podCreationTimestamp="2025-12-09 17:38:36 +0000 UTC" firstStartedPulling="2025-12-09 17:38:37.671351391 +0000 UTC m=+2514.059525211" lastFinishedPulling="2025-12-09 17:38:38.353695962 +0000 UTC m=+2514.741869782" observedRunningTime="2025-12-09 17:38:38.738277158 +0000 UTC m=+2515.126451068" watchObservedRunningTime="2025-12-09 17:38:46.03455431 +0000 UTC m=+2522.422728130" Dec 09 17:38:46 crc kubenswrapper[4954]: I1209 17:38:46.040733 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-z88n9"] Dec 09 17:38:46 crc kubenswrapper[4954]: I1209 17:38:46.052976 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-bce8-account-create-update-lhxt6"] Dec 09 17:38:46 crc kubenswrapper[4954]: I1209 17:38:46.062668 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-kdt2k"] Dec 09 17:38:46 crc kubenswrapper[4954]: I1209 17:38:46.072542 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-bce8-account-create-update-lhxt6"] Dec 09 17:38:46 crc kubenswrapper[4954]: I1209 17:38:46.082562 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-z88n9"] Dec 09 17:38:46 crc kubenswrapper[4954]: I1209 17:38:46.093050 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-kdt2k"] Dec 09 17:38:46 crc kubenswrapper[4954]: I1209 17:38:46.134668 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10f8b023-726d-443f-b728-eaa574d45bab" path="/var/lib/kubelet/pods/10f8b023-726d-443f-b728-eaa574d45bab/volumes" Dec 09 17:38:46 crc kubenswrapper[4954]: I1209 17:38:46.135635 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41a8d52b-549b-42d1-91bf-e7e4567affb3" path="/var/lib/kubelet/pods/41a8d52b-549b-42d1-91bf-e7e4567affb3/volumes" Dec 09 17:38:46 crc kubenswrapper[4954]: I1209 17:38:46.136489 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f599c00a-fbf5-46cb-9304-f5554664eb3e" path="/var/lib/kubelet/pods/f599c00a-fbf5-46cb-9304-f5554664eb3e/volumes" Dec 09 17:38:47 crc kubenswrapper[4954]: I1209 17:38:47.035177 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-c84cd"] Dec 09 17:38:47 crc kubenswrapper[4954]: I1209 17:38:47.049129 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-a4ac-account-create-update-qprph"] Dec 09 17:38:47 crc kubenswrapper[4954]: I1209 17:38:47.062173 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-c84cd"] Dec 09 17:38:47 crc kubenswrapper[4954]: I1209 17:38:47.085882 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-602a-account-create-update-892sk"] Dec 09 17:38:47 crc kubenswrapper[4954]: I1209 17:38:47.096955 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-602a-account-create-update-892sk"] Dec 09 17:38:47 crc kubenswrapper[4954]: I1209 17:38:47.107111 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-a4ac-account-create-update-qprph"] Dec 09 17:38:48 crc kubenswrapper[4954]: E1209 17:38:48.125881 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:38:48 crc kubenswrapper[4954]: I1209 17:38:48.137232 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1683bbec-ff24-4d14-aed8-cfb49b328355" path="/var/lib/kubelet/pods/1683bbec-ff24-4d14-aed8-cfb49b328355/volumes" Dec 09 17:38:48 crc kubenswrapper[4954]: I1209 17:38:48.138291 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="588cbebb-ba39-4145-9c5e-865f77ee92cb" path="/var/lib/kubelet/pods/588cbebb-ba39-4145-9c5e-865f77ee92cb/volumes" Dec 09 17:38:48 crc kubenswrapper[4954]: I1209 17:38:48.139411 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d295cb-5d01-4014-abbe-efc80d8259f8" path="/var/lib/kubelet/pods/d7d295cb-5d01-4014-abbe-efc80d8259f8/volumes" Dec 09 17:38:52 crc kubenswrapper[4954]: I1209 17:38:52.120332 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:38:52 crc kubenswrapper[4954]: E1209 17:38:52.121276 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:38:59 crc kubenswrapper[4954]: E1209 17:38:59.122330 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:38:59 crc kubenswrapper[4954]: E1209 17:38:59.122419 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:39:04 crc kubenswrapper[4954]: I1209 17:39:04.169626 4954 scope.go:117] "RemoveContainer" containerID="3ef13b2db96f3a68ca5e0d44d2301257ab37079cdd3f1aaa08a84bb45e3733d8" Dec 09 17:39:04 crc kubenswrapper[4954]: I1209 17:39:04.207279 4954 scope.go:117] "RemoveContainer" containerID="019944e8bf6e1f3b40a622ec027ff5dd4864764b4b0a13313c3f50e00cb28fd5" Dec 09 17:39:04 crc kubenswrapper[4954]: I1209 17:39:04.250721 4954 scope.go:117] "RemoveContainer" containerID="dbde0662e80db3b450369d2715e456152bbd324324f5b0e040fae5521791b31c" Dec 09 17:39:04 crc kubenswrapper[4954]: I1209 17:39:04.307166 4954 scope.go:117] "RemoveContainer" containerID="16c693de493d33217dab79116f369ab5b07d38b359ecc03353517560d4a21f6b" Dec 09 17:39:04 crc kubenswrapper[4954]: I1209 17:39:04.366150 4954 scope.go:117] "RemoveContainer" containerID="b612b367e033820f2f8473c0faf858404b5431b61e6e8f0a12445111edf8be64" Dec 09 17:39:04 crc kubenswrapper[4954]: I1209 17:39:04.406528 4954 scope.go:117] "RemoveContainer" containerID="013ac2deeccccad2e28d9d3ccdba7f6af5648419987820c5543ba5893f5717c2" Dec 09 17:39:06 crc kubenswrapper[4954]: I1209 17:39:06.120053 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:39:06 crc kubenswrapper[4954]: E1209 17:39:06.120620 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:39:11 crc kubenswrapper[4954]: E1209 17:39:11.122635 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:39:12 crc kubenswrapper[4954]: E1209 17:39:12.123344 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:39:16 crc kubenswrapper[4954]: I1209 17:39:16.054285 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ln5zl"] Dec 09 17:39:16 crc kubenswrapper[4954]: I1209 17:39:16.063160 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ln5zl"] Dec 09 17:39:16 crc kubenswrapper[4954]: I1209 17:39:16.132735 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24d42cd5-dc37-4675-87b9-91dcb7391eed" path="/var/lib/kubelet/pods/24d42cd5-dc37-4675-87b9-91dcb7391eed/volumes" Dec 09 17:39:20 crc kubenswrapper[4954]: I1209 17:39:20.120505 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:39:20 crc kubenswrapper[4954]: E1209 17:39:20.121365 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:39:25 crc kubenswrapper[4954]: E1209 17:39:25.122917 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:39:26 crc kubenswrapper[4954]: E1209 17:39:26.121495 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:39:27 crc kubenswrapper[4954]: I1209 17:39:27.700902 4954 generic.go:334] "Generic (PLEG): container finished" podID="b6a4b196-e34d-4b20-9915-f290a4dac804" containerID="32f0c146d7dd24e54a8e0fc654d88db6d1ecbdfce7b2c06834e1b3d8d5114754" exitCode=0 Dec 09 17:39:27 crc kubenswrapper[4954]: I1209 17:39:27.701132 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" event={"ID":"b6a4b196-e34d-4b20-9915-f290a4dac804","Type":"ContainerDied","Data":"32f0c146d7dd24e54a8e0fc654d88db6d1ecbdfce7b2c06834e1b3d8d5114754"} Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.187323 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.367571 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-ssh-key\") pod \"b6a4b196-e34d-4b20-9915-f290a4dac804\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.367897 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhvrm\" (UniqueName: \"kubernetes.io/projected/b6a4b196-e34d-4b20-9915-f290a4dac804-kube-api-access-hhvrm\") pod \"b6a4b196-e34d-4b20-9915-f290a4dac804\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.367974 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-inventory\") pod \"b6a4b196-e34d-4b20-9915-f290a4dac804\" (UID: \"b6a4b196-e34d-4b20-9915-f290a4dac804\") " Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.373943 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6a4b196-e34d-4b20-9915-f290a4dac804-kube-api-access-hhvrm" (OuterVolumeSpecName: "kube-api-access-hhvrm") pod "b6a4b196-e34d-4b20-9915-f290a4dac804" (UID: "b6a4b196-e34d-4b20-9915-f290a4dac804"). InnerVolumeSpecName "kube-api-access-hhvrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.398460 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b6a4b196-e34d-4b20-9915-f290a4dac804" (UID: "b6a4b196-e34d-4b20-9915-f290a4dac804"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.403191 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-inventory" (OuterVolumeSpecName: "inventory") pod "b6a4b196-e34d-4b20-9915-f290a4dac804" (UID: "b6a4b196-e34d-4b20-9915-f290a4dac804"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.470546 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.470645 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhvrm\" (UniqueName: \"kubernetes.io/projected/b6a4b196-e34d-4b20-9915-f290a4dac804-kube-api-access-hhvrm\") on node \"crc\" DevicePath \"\"" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.470670 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6a4b196-e34d-4b20-9915-f290a4dac804-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.721856 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" event={"ID":"b6a4b196-e34d-4b20-9915-f290a4dac804","Type":"ContainerDied","Data":"0cb35adfae29332c69edf7df0999f2b187d74c1357ecb951230b2320f8d207f6"} Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.721905 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cb35adfae29332c69edf7df0999f2b187d74c1357ecb951230b2320f8d207f6" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.722003 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.828920 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2w25t"] Dec 09 17:39:29 crc kubenswrapper[4954]: E1209 17:39:29.829876 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6a4b196-e34d-4b20-9915-f290a4dac804" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.829997 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6a4b196-e34d-4b20-9915-f290a4dac804" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.830271 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6a4b196-e34d-4b20-9915-f290a4dac804" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.831107 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.834048 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.834230 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.834363 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.834560 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.847265 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2w25t"] Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.980238 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2w25t\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.980672 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2w25t\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:29 crc kubenswrapper[4954]: I1209 17:39:29.980783 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkv4m\" (UniqueName: \"kubernetes.io/projected/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-kube-api-access-mkv4m\") pod \"ssh-known-hosts-edpm-deployment-2w25t\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:30 crc kubenswrapper[4954]: I1209 17:39:30.082999 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2w25t\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:30 crc kubenswrapper[4954]: I1209 17:39:30.083097 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2w25t\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:30 crc kubenswrapper[4954]: I1209 17:39:30.083248 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkv4m\" (UniqueName: \"kubernetes.io/projected/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-kube-api-access-mkv4m\") pod \"ssh-known-hosts-edpm-deployment-2w25t\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:30 crc kubenswrapper[4954]: I1209 17:39:30.088080 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2w25t\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:30 crc kubenswrapper[4954]: I1209 17:39:30.089127 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2w25t\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:30 crc kubenswrapper[4954]: I1209 17:39:30.100256 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkv4m\" (UniqueName: \"kubernetes.io/projected/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-kube-api-access-mkv4m\") pod \"ssh-known-hosts-edpm-deployment-2w25t\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:30 crc kubenswrapper[4954]: I1209 17:39:30.150135 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:30 crc kubenswrapper[4954]: I1209 17:39:30.652060 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2w25t"] Dec 09 17:39:30 crc kubenswrapper[4954]: I1209 17:39:30.731712 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" event={"ID":"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1","Type":"ContainerStarted","Data":"b84fa3e7f90aa1abed6b4da3ececc4db36de6dac0ac7610a7392f3895e035313"} Dec 09 17:39:31 crc kubenswrapper[4954]: I1209 17:39:31.741950 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" event={"ID":"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1","Type":"ContainerStarted","Data":"8fca2e32ee6900faba8fc40d1cb481124bcda0abd0a187b19dcd6b273dd37b1d"} Dec 09 17:39:31 crc kubenswrapper[4954]: I1209 17:39:31.765490 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" podStartSLOduration=2.263149835 podStartE2EDuration="2.765462185s" podCreationTimestamp="2025-12-09 17:39:29 +0000 UTC" firstStartedPulling="2025-12-09 17:39:30.662534025 +0000 UTC m=+2567.050707845" lastFinishedPulling="2025-12-09 17:39:31.164846375 +0000 UTC m=+2567.553020195" observedRunningTime="2025-12-09 17:39:31.758102864 +0000 UTC m=+2568.146276694" watchObservedRunningTime="2025-12-09 17:39:31.765462185 +0000 UTC m=+2568.153636005" Dec 09 17:39:32 crc kubenswrapper[4954]: I1209 17:39:32.120846 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:39:32 crc kubenswrapper[4954]: E1209 17:39:32.121180 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:39:33 crc kubenswrapper[4954]: I1209 17:39:33.029107 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-6jg6z"] Dec 09 17:39:33 crc kubenswrapper[4954]: I1209 17:39:33.040299 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-6jg6z"] Dec 09 17:39:34 crc kubenswrapper[4954]: I1209 17:39:34.029553 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-52c6-account-create-update-gcs5d"] Dec 09 17:39:34 crc kubenswrapper[4954]: I1209 17:39:34.040564 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-52c6-account-create-update-gcs5d"] Dec 09 17:39:34 crc kubenswrapper[4954]: I1209 17:39:34.131613 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="216180c8-daea-4b80-8d94-659e92904a12" path="/var/lib/kubelet/pods/216180c8-daea-4b80-8d94-659e92904a12/volumes" Dec 09 17:39:34 crc kubenswrapper[4954]: I1209 17:39:34.132363 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="436a9623-3c31-40be-a3bc-e4e3364393ac" path="/var/lib/kubelet/pods/436a9623-3c31-40be-a3bc-e4e3364393ac/volumes" Dec 09 17:39:36 crc kubenswrapper[4954]: E1209 17:39:36.123677 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:39:38 crc kubenswrapper[4954]: I1209 17:39:38.802628 4954 generic.go:334] "Generic (PLEG): container finished" podID="b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1" containerID="8fca2e32ee6900faba8fc40d1cb481124bcda0abd0a187b19dcd6b273dd37b1d" exitCode=0 Dec 09 17:39:38 crc kubenswrapper[4954]: I1209 17:39:38.802830 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" event={"ID":"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1","Type":"ContainerDied","Data":"8fca2e32ee6900faba8fc40d1cb481124bcda0abd0a187b19dcd6b273dd37b1d"} Dec 09 17:39:40 crc kubenswrapper[4954]: E1209 17:39:40.122869 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.244963 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.405506 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-inventory-0\") pod \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.406026 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-ssh-key-openstack-edpm-ipam\") pod \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.406136 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkv4m\" (UniqueName: \"kubernetes.io/projected/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-kube-api-access-mkv4m\") pod \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\" (UID: \"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1\") " Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.410924 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-kube-api-access-mkv4m" (OuterVolumeSpecName: "kube-api-access-mkv4m") pod "b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1" (UID: "b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1"). InnerVolumeSpecName "kube-api-access-mkv4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.435218 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1" (UID: "b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.439796 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1" (UID: "b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.508899 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.508952 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkv4m\" (UniqueName: \"kubernetes.io/projected/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-kube-api-access-mkv4m\") on node \"crc\" DevicePath \"\"" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.508965 4954 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.823226 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" event={"ID":"b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1","Type":"ContainerDied","Data":"b84fa3e7f90aa1abed6b4da3ececc4db36de6dac0ac7610a7392f3895e035313"} Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.823279 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b84fa3e7f90aa1abed6b4da3ececc4db36de6dac0ac7610a7392f3895e035313" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.823297 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2w25t" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.896781 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4"] Dec 09 17:39:40 crc kubenswrapper[4954]: E1209 17:39:40.897303 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1" containerName="ssh-known-hosts-edpm-deployment" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.897321 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1" containerName="ssh-known-hosts-edpm-deployment" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.897610 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1" containerName="ssh-known-hosts-edpm-deployment" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.898433 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.900709 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.900902 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.901338 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.904516 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:39:40 crc kubenswrapper[4954]: I1209 17:39:40.906961 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4"] Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.018988 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rvnx4\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.019369 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rvnx4\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.019733 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k8qc\" (UniqueName: \"kubernetes.io/projected/f7031e05-5441-47fa-84f5-f35fada4a4ba-kube-api-access-4k8qc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rvnx4\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.121502 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rvnx4\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.121692 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k8qc\" (UniqueName: \"kubernetes.io/projected/f7031e05-5441-47fa-84f5-f35fada4a4ba-kube-api-access-4k8qc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rvnx4\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.121796 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rvnx4\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.135405 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rvnx4\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.136280 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rvnx4\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.138016 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k8qc\" (UniqueName: \"kubernetes.io/projected/f7031e05-5441-47fa-84f5-f35fada4a4ba-kube-api-access-4k8qc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-rvnx4\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.224697 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.762560 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4"] Dec 09 17:39:41 crc kubenswrapper[4954]: W1209 17:39:41.765846 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7031e05_5441_47fa_84f5_f35fada4a4ba.slice/crio-a44908e6cfd5cddda512ed0cefe6ecdeeb958e16d1c476d542aff9d8c586fdd3 WatchSource:0}: Error finding container a44908e6cfd5cddda512ed0cefe6ecdeeb958e16d1c476d542aff9d8c586fdd3: Status 404 returned error can't find the container with id a44908e6cfd5cddda512ed0cefe6ecdeeb958e16d1c476d542aff9d8c586fdd3 Dec 09 17:39:41 crc kubenswrapper[4954]: I1209 17:39:41.835731 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" event={"ID":"f7031e05-5441-47fa-84f5-f35fada4a4ba","Type":"ContainerStarted","Data":"a44908e6cfd5cddda512ed0cefe6ecdeeb958e16d1c476d542aff9d8c586fdd3"} Dec 09 17:39:42 crc kubenswrapper[4954]: I1209 17:39:42.849216 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" event={"ID":"f7031e05-5441-47fa-84f5-f35fada4a4ba","Type":"ContainerStarted","Data":"e38b47a8e1a5b5063a4684971bc43e72cd98cb1df074b58181735c82b76c7092"} Dec 09 17:39:42 crc kubenswrapper[4954]: I1209 17:39:42.878898 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" podStartSLOduration=2.497370016 podStartE2EDuration="2.878875626s" podCreationTimestamp="2025-12-09 17:39:40 +0000 UTC" firstStartedPulling="2025-12-09 17:39:41.768812401 +0000 UTC m=+2578.156986221" lastFinishedPulling="2025-12-09 17:39:42.150318011 +0000 UTC m=+2578.538491831" observedRunningTime="2025-12-09 17:39:42.870967747 +0000 UTC m=+2579.259141587" watchObservedRunningTime="2025-12-09 17:39:42.878875626 +0000 UTC m=+2579.267049446" Dec 09 17:39:46 crc kubenswrapper[4954]: I1209 17:39:46.064795 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-4gb2h"] Dec 09 17:39:46 crc kubenswrapper[4954]: I1209 17:39:46.083283 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-blmj4"] Dec 09 17:39:46 crc kubenswrapper[4954]: I1209 17:39:46.104294 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-4gb2h"] Dec 09 17:39:46 crc kubenswrapper[4954]: I1209 17:39:46.113901 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-blmj4"] Dec 09 17:39:46 crc kubenswrapper[4954]: I1209 17:39:46.121229 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:39:46 crc kubenswrapper[4954]: E1209 17:39:46.121851 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:39:46 crc kubenswrapper[4954]: I1209 17:39:46.136721 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b263b3a-d693-45c2-8d18-ff0170ee5851" path="/var/lib/kubelet/pods/0b263b3a-d693-45c2-8d18-ff0170ee5851/volumes" Dec 09 17:39:46 crc kubenswrapper[4954]: I1209 17:39:46.137509 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="884b0d29-b0dd-48bf-bb02-51cbb4503b41" path="/var/lib/kubelet/pods/884b0d29-b0dd-48bf-bb02-51cbb4503b41/volumes" Dec 09 17:39:50 crc kubenswrapper[4954]: E1209 17:39:50.123083 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:39:50 crc kubenswrapper[4954]: I1209 17:39:50.924025 4954 generic.go:334] "Generic (PLEG): container finished" podID="f7031e05-5441-47fa-84f5-f35fada4a4ba" containerID="e38b47a8e1a5b5063a4684971bc43e72cd98cb1df074b58181735c82b76c7092" exitCode=0 Dec 09 17:39:50 crc kubenswrapper[4954]: I1209 17:39:50.924120 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" event={"ID":"f7031e05-5441-47fa-84f5-f35fada4a4ba","Type":"ContainerDied","Data":"e38b47a8e1a5b5063a4684971bc43e72cd98cb1df074b58181735c82b76c7092"} Dec 09 17:39:51 crc kubenswrapper[4954]: I1209 17:39:51.028221 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-6q2z2"] Dec 09 17:39:51 crc kubenswrapper[4954]: I1209 17:39:51.040968 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-6q2z2"] Dec 09 17:39:52 crc kubenswrapper[4954]: E1209 17:39:52.126158 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.133369 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f107a10b-0481-41b9-bcfa-7e5bef86ca1a" path="/var/lib/kubelet/pods/f107a10b-0481-41b9-bcfa-7e5bef86ca1a/volumes" Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.378253 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.460959 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-inventory\") pod \"f7031e05-5441-47fa-84f5-f35fada4a4ba\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.461302 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-ssh-key\") pod \"f7031e05-5441-47fa-84f5-f35fada4a4ba\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.461374 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4k8qc\" (UniqueName: \"kubernetes.io/projected/f7031e05-5441-47fa-84f5-f35fada4a4ba-kube-api-access-4k8qc\") pod \"f7031e05-5441-47fa-84f5-f35fada4a4ba\" (UID: \"f7031e05-5441-47fa-84f5-f35fada4a4ba\") " Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.466929 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7031e05-5441-47fa-84f5-f35fada4a4ba-kube-api-access-4k8qc" (OuterVolumeSpecName: "kube-api-access-4k8qc") pod "f7031e05-5441-47fa-84f5-f35fada4a4ba" (UID: "f7031e05-5441-47fa-84f5-f35fada4a4ba"). InnerVolumeSpecName "kube-api-access-4k8qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.500632 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-inventory" (OuterVolumeSpecName: "inventory") pod "f7031e05-5441-47fa-84f5-f35fada4a4ba" (UID: "f7031e05-5441-47fa-84f5-f35fada4a4ba"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.501422 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f7031e05-5441-47fa-84f5-f35fada4a4ba" (UID: "f7031e05-5441-47fa-84f5-f35fada4a4ba"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.564552 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.564955 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4k8qc\" (UniqueName: \"kubernetes.io/projected/f7031e05-5441-47fa-84f5-f35fada4a4ba-kube-api-access-4k8qc\") on node \"crc\" DevicePath \"\"" Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.564970 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7031e05-5441-47fa-84f5-f35fada4a4ba-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.949510 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" event={"ID":"f7031e05-5441-47fa-84f5-f35fada4a4ba","Type":"ContainerDied","Data":"a44908e6cfd5cddda512ed0cefe6ecdeeb958e16d1c476d542aff9d8c586fdd3"} Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.949554 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-rvnx4" Dec 09 17:39:52 crc kubenswrapper[4954]: I1209 17:39:52.949557 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a44908e6cfd5cddda512ed0cefe6ecdeeb958e16d1c476d542aff9d8c586fdd3" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.012906 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8"] Dec 09 17:39:53 crc kubenswrapper[4954]: E1209 17:39:53.013420 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7031e05-5441-47fa-84f5-f35fada4a4ba" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.013438 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7031e05-5441-47fa-84f5-f35fada4a4ba" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.013642 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7031e05-5441-47fa-84f5-f35fada4a4ba" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.014372 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.016663 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.017249 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.017248 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.019046 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.029170 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8"] Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.075328 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.075690 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8ddr\" (UniqueName: \"kubernetes.io/projected/a7681334-f1d0-409b-a4e0-742786731cda-kube-api-access-r8ddr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.075894 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.179297 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.180070 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.180610 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8ddr\" (UniqueName: \"kubernetes.io/projected/a7681334-f1d0-409b-a4e0-742786731cda-kube-api-access-r8ddr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.183389 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.189486 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.198920 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8ddr\" (UniqueName: \"kubernetes.io/projected/a7681334-f1d0-409b-a4e0-742786731cda-kube-api-access-r8ddr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.330254 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.839367 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8"] Dec 09 17:39:53 crc kubenswrapper[4954]: W1209 17:39:53.840827 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7681334_f1d0_409b_a4e0_742786731cda.slice/crio-b1fd4386f7a3a06828e5fdc007ff5a6ae8276e13a47c39526ebb9985f968a0b7 WatchSource:0}: Error finding container b1fd4386f7a3a06828e5fdc007ff5a6ae8276e13a47c39526ebb9985f968a0b7: Status 404 returned error can't find the container with id b1fd4386f7a3a06828e5fdc007ff5a6ae8276e13a47c39526ebb9985f968a0b7 Dec 09 17:39:53 crc kubenswrapper[4954]: I1209 17:39:53.974630 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" event={"ID":"a7681334-f1d0-409b-a4e0-742786731cda","Type":"ContainerStarted","Data":"b1fd4386f7a3a06828e5fdc007ff5a6ae8276e13a47c39526ebb9985f968a0b7"} Dec 09 17:39:54 crc kubenswrapper[4954]: I1209 17:39:54.987478 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" event={"ID":"a7681334-f1d0-409b-a4e0-742786731cda","Type":"ContainerStarted","Data":"899fd796cf216611632f9362f20b821f3256a68e349e38b4841af0b4f9efcde1"} Dec 09 17:39:55 crc kubenswrapper[4954]: I1209 17:39:55.011060 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" podStartSLOduration=2.455073761 podStartE2EDuration="3.011041478s" podCreationTimestamp="2025-12-09 17:39:52 +0000 UTC" firstStartedPulling="2025-12-09 17:39:53.843338641 +0000 UTC m=+2590.231512461" lastFinishedPulling="2025-12-09 17:39:54.399306358 +0000 UTC m=+2590.787480178" observedRunningTime="2025-12-09 17:39:55.004870114 +0000 UTC m=+2591.393043964" watchObservedRunningTime="2025-12-09 17:39:55.011041478 +0000 UTC m=+2591.399215298" Dec 09 17:39:57 crc kubenswrapper[4954]: I1209 17:39:57.120403 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:39:57 crc kubenswrapper[4954]: E1209 17:39:57.121028 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:40:03 crc kubenswrapper[4954]: E1209 17:40:03.124244 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:40:04 crc kubenswrapper[4954]: I1209 17:40:04.081641 4954 generic.go:334] "Generic (PLEG): container finished" podID="a7681334-f1d0-409b-a4e0-742786731cda" containerID="899fd796cf216611632f9362f20b821f3256a68e349e38b4841af0b4f9efcde1" exitCode=0 Dec 09 17:40:04 crc kubenswrapper[4954]: I1209 17:40:04.081686 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" event={"ID":"a7681334-f1d0-409b-a4e0-742786731cda","Type":"ContainerDied","Data":"899fd796cf216611632f9362f20b821f3256a68e349e38b4841af0b4f9efcde1"} Dec 09 17:40:04 crc kubenswrapper[4954]: E1209 17:40:04.134700 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:40:04 crc kubenswrapper[4954]: I1209 17:40:04.572443 4954 scope.go:117] "RemoveContainer" containerID="f38e068ac82370a127fa322b93040f4b78057d4ffd9da46a8d332807fbb35d2a" Dec 09 17:40:04 crc kubenswrapper[4954]: I1209 17:40:04.625899 4954 scope.go:117] "RemoveContainer" containerID="8b9de092f02f4523780f2eaf864710b98a587bde05d42ed00dd49e8e9d6a0a48" Dec 09 17:40:04 crc kubenswrapper[4954]: I1209 17:40:04.668419 4954 scope.go:117] "RemoveContainer" containerID="9bd5edc6798b8894974430538ce7f67ab22bb12c2d70de46d5e42f01fe505ae2" Dec 09 17:40:04 crc kubenswrapper[4954]: I1209 17:40:04.704323 4954 scope.go:117] "RemoveContainer" containerID="289fc5c5a373dc0fbb22475c3925fb11dc75b13b2f17fb7b28cdf313f453841a" Dec 09 17:40:04 crc kubenswrapper[4954]: I1209 17:40:04.775360 4954 scope.go:117] "RemoveContainer" containerID="b399558fc9a93c7bdefe8831a86a84e5442ffe70b650a46d464c11a06da8c3bc" Dec 09 17:40:04 crc kubenswrapper[4954]: I1209 17:40:04.803085 4954 scope.go:117] "RemoveContainer" containerID="c3d8b0f1fea79e6696f51c27a8d5c41ab2f966b35471565ddc1efa2bc808e962" Dec 09 17:40:05 crc kubenswrapper[4954]: I1209 17:40:05.440856 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:40:05 crc kubenswrapper[4954]: I1209 17:40:05.541955 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-inventory\") pod \"a7681334-f1d0-409b-a4e0-742786731cda\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " Dec 09 17:40:05 crc kubenswrapper[4954]: I1209 17:40:05.542044 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-ssh-key\") pod \"a7681334-f1d0-409b-a4e0-742786731cda\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " Dec 09 17:40:05 crc kubenswrapper[4954]: I1209 17:40:05.542211 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8ddr\" (UniqueName: \"kubernetes.io/projected/a7681334-f1d0-409b-a4e0-742786731cda-kube-api-access-r8ddr\") pod \"a7681334-f1d0-409b-a4e0-742786731cda\" (UID: \"a7681334-f1d0-409b-a4e0-742786731cda\") " Dec 09 17:40:05 crc kubenswrapper[4954]: I1209 17:40:05.547108 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7681334-f1d0-409b-a4e0-742786731cda-kube-api-access-r8ddr" (OuterVolumeSpecName: "kube-api-access-r8ddr") pod "a7681334-f1d0-409b-a4e0-742786731cda" (UID: "a7681334-f1d0-409b-a4e0-742786731cda"). InnerVolumeSpecName "kube-api-access-r8ddr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:40:05 crc kubenswrapper[4954]: I1209 17:40:05.571102 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a7681334-f1d0-409b-a4e0-742786731cda" (UID: "a7681334-f1d0-409b-a4e0-742786731cda"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:40:05 crc kubenswrapper[4954]: I1209 17:40:05.571843 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-inventory" (OuterVolumeSpecName: "inventory") pod "a7681334-f1d0-409b-a4e0-742786731cda" (UID: "a7681334-f1d0-409b-a4e0-742786731cda"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:40:05 crc kubenswrapper[4954]: I1209 17:40:05.645567 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:05 crc kubenswrapper[4954]: I1209 17:40:05.645633 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8ddr\" (UniqueName: \"kubernetes.io/projected/a7681334-f1d0-409b-a4e0-742786731cda-kube-api-access-r8ddr\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:05 crc kubenswrapper[4954]: I1209 17:40:05.645651 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7681334-f1d0-409b-a4e0-742786731cda-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.100478 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" event={"ID":"a7681334-f1d0-409b-a4e0-742786731cda","Type":"ContainerDied","Data":"b1fd4386f7a3a06828e5fdc007ff5a6ae8276e13a47c39526ebb9985f968a0b7"} Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.100518 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1fd4386f7a3a06828e5fdc007ff5a6ae8276e13a47c39526ebb9985f968a0b7" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.100557 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.191971 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd"] Dec 09 17:40:06 crc kubenswrapper[4954]: E1209 17:40:06.192442 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7681334-f1d0-409b-a4e0-742786731cda" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.192461 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7681334-f1d0-409b-a4e0-742786731cda" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.192715 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7681334-f1d0-409b-a4e0-742786731cda" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.193679 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.199965 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.200053 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.200113 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.200128 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.200166 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.200220 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.200228 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.200467 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.211470 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd"] Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257350 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257420 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257449 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257487 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257518 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257581 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257613 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257647 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257675 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257696 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257725 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257768 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdwrp\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-kube-api-access-vdwrp\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.257787 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.359885 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.359937 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.359983 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.360027 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.360094 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.360114 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.360145 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.360173 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.360196 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.360252 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.360300 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdwrp\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-kube-api-access-vdwrp\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.360319 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.360352 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.364323 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.364758 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.364926 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.365457 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.365469 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.365656 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.366509 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.366574 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.366719 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.366746 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.364690 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.376382 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.377495 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdwrp\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-kube-api-access-vdwrp\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:06 crc kubenswrapper[4954]: I1209 17:40:06.518284 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:07 crc kubenswrapper[4954]: I1209 17:40:07.044052 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd"] Dec 09 17:40:07 crc kubenswrapper[4954]: I1209 17:40:07.111512 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" event={"ID":"2251824f-c169-44ff-afeb-8649fb57c39f","Type":"ContainerStarted","Data":"a9baa38159f628b07659a4bc2c687321c95fcf495603fed752e93f32753a8440"} Dec 09 17:40:08 crc kubenswrapper[4954]: I1209 17:40:08.124380 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" event={"ID":"2251824f-c169-44ff-afeb-8649fb57c39f","Type":"ContainerStarted","Data":"ce82736bcf849148a72e18b4c3784553b468726c4ed4ba488f9c5acc6abdac13"} Dec 09 17:40:08 crc kubenswrapper[4954]: I1209 17:40:08.152730 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" podStartSLOduration=1.627429902 podStartE2EDuration="2.152712723s" podCreationTimestamp="2025-12-09 17:40:06 +0000 UTC" firstStartedPulling="2025-12-09 17:40:07.056573117 +0000 UTC m=+2603.444746937" lastFinishedPulling="2025-12-09 17:40:07.581855938 +0000 UTC m=+2603.970029758" observedRunningTime="2025-12-09 17:40:08.143037688 +0000 UTC m=+2604.531211508" watchObservedRunningTime="2025-12-09 17:40:08.152712723 +0000 UTC m=+2604.540886553" Dec 09 17:40:11 crc kubenswrapper[4954]: I1209 17:40:11.120081 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:40:11 crc kubenswrapper[4954]: E1209 17:40:11.120625 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:40:15 crc kubenswrapper[4954]: E1209 17:40:15.122109 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:40:17 crc kubenswrapper[4954]: E1209 17:40:17.123792 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:40:22 crc kubenswrapper[4954]: I1209 17:40:22.120323 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:40:22 crc kubenswrapper[4954]: E1209 17:40:22.121218 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:40:28 crc kubenswrapper[4954]: I1209 17:40:28.044573 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-4bl8c"] Dec 09 17:40:28 crc kubenswrapper[4954]: I1209 17:40:28.055398 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-4bl8c"] Dec 09 17:40:28 crc kubenswrapper[4954]: E1209 17:40:28.123564 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:40:28 crc kubenswrapper[4954]: I1209 17:40:28.137728 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="961eb076-69ef-4160-80db-3b44f94f5a73" path="/var/lib/kubelet/pods/961eb076-69ef-4160-80db-3b44f94f5a73/volumes" Dec 09 17:40:32 crc kubenswrapper[4954]: E1209 17:40:32.123543 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:40:33 crc kubenswrapper[4954]: I1209 17:40:33.120631 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:40:33 crc kubenswrapper[4954]: E1209 17:40:33.121226 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:40:40 crc kubenswrapper[4954]: E1209 17:40:40.124304 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:40:42 crc kubenswrapper[4954]: I1209 17:40:42.432133 4954 generic.go:334] "Generic (PLEG): container finished" podID="2251824f-c169-44ff-afeb-8649fb57c39f" containerID="ce82736bcf849148a72e18b4c3784553b468726c4ed4ba488f9c5acc6abdac13" exitCode=0 Dec 09 17:40:42 crc kubenswrapper[4954]: I1209 17:40:42.432392 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" event={"ID":"2251824f-c169-44ff-afeb-8649fb57c39f","Type":"ContainerDied","Data":"ce82736bcf849148a72e18b4c3784553b468726c4ed4ba488f9c5acc6abdac13"} Dec 09 17:40:43 crc kubenswrapper[4954]: I1209 17:40:43.960539 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.021617 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-inventory\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.021682 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-repo-setup-combined-ca-bundle\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.021712 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdwrp\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-kube-api-access-vdwrp\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.021743 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-combined-ca-bundle\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.021766 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-power-monitoring-combined-ca-bundle\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.021797 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.021874 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.021975 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ssh-key\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.022069 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.022119 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.022141 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-libvirt-combined-ca-bundle\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.022195 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ovn-combined-ca-bundle\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.022311 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-bootstrap-combined-ca-bundle\") pod \"2251824f-c169-44ff-afeb-8649fb57c39f\" (UID: \"2251824f-c169-44ff-afeb-8649fb57c39f\") " Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.030486 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.030645 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-kube-api-access-vdwrp" (OuterVolumeSpecName: "kube-api-access-vdwrp") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "kube-api-access-vdwrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.031667 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.037278 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.042032 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.043433 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.044215 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.044454 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.046503 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.048037 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.049214 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.068512 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.087012 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-inventory" (OuterVolumeSpecName: "inventory") pod "2251824f-c169-44ff-afeb-8649fb57c39f" (UID: "2251824f-c169-44ff-afeb-8649fb57c39f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127274 4954 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127319 4954 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127336 4954 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127350 4954 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127362 4954 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127374 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127386 4954 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127417 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdwrp\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-kube-api-access-vdwrp\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127430 4954 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127441 4954 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127452 4954 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127465 4954 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2251824f-c169-44ff-afeb-8649fb57c39f-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.127477 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2251824f-c169-44ff-afeb-8649fb57c39f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.454227 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" event={"ID":"2251824f-c169-44ff-afeb-8649fb57c39f","Type":"ContainerDied","Data":"a9baa38159f628b07659a4bc2c687321c95fcf495603fed752e93f32753a8440"} Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.454490 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9baa38159f628b07659a4bc2c687321c95fcf495603fed752e93f32753a8440" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.454323 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.463768 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xg6qk"] Dec 09 17:40:44 crc kubenswrapper[4954]: E1209 17:40:44.464456 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2251824f-c169-44ff-afeb-8649fb57c39f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.464484 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2251824f-c169-44ff-afeb-8649fb57c39f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.464756 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="2251824f-c169-44ff-afeb-8649fb57c39f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.466797 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.503671 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xg6qk"] Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.537315 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-catalog-content\") pod \"community-operators-xg6qk\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.537483 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-utilities\") pod \"community-operators-xg6qk\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.537513 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nthjj\" (UniqueName: \"kubernetes.io/projected/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-kube-api-access-nthjj\") pod \"community-operators-xg6qk\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.587063 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl"] Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.588952 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.593552 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.594002 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.593702 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.593935 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.602018 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl"] Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.602242 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.640459 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pzn9\" (UniqueName: \"kubernetes.io/projected/a20b6b58-cb72-432d-8ec8-7c7f93265d36-kube-api-access-2pzn9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.640666 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.640728 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-catalog-content\") pod \"community-operators-xg6qk\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.640946 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.640983 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-utilities\") pod \"community-operators-xg6qk\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.641017 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nthjj\" (UniqueName: \"kubernetes.io/projected/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-kube-api-access-nthjj\") pod \"community-operators-xg6qk\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.641118 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.641207 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.641856 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-catalog-content\") pod \"community-operators-xg6qk\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.642055 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-utilities\") pod \"community-operators-xg6qk\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.662840 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nthjj\" (UniqueName: \"kubernetes.io/projected/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-kube-api-access-nthjj\") pod \"community-operators-xg6qk\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.743020 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.743108 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.743158 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.743236 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pzn9\" (UniqueName: \"kubernetes.io/projected/a20b6b58-cb72-432d-8ec8-7c7f93265d36-kube-api-access-2pzn9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.743293 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.744173 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.746961 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.746974 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.748120 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.761699 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pzn9\" (UniqueName: \"kubernetes.io/projected/a20b6b58-cb72-432d-8ec8-7c7f93265d36-kube-api-access-2pzn9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jt7gl\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.829401 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:44 crc kubenswrapper[4954]: I1209 17:40:44.911932 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:40:45 crc kubenswrapper[4954]: E1209 17:40:45.122557 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:40:45 crc kubenswrapper[4954]: I1209 17:40:45.314264 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xg6qk"] Dec 09 17:40:45 crc kubenswrapper[4954]: W1209 17:40:45.324098 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d9f331b_1de4_4d06_8caa_7115a9e1c10e.slice/crio-81f31016c1cba082b07d34c2d65a70faf6f6b3779acaefcaaf756aa2037d04fb WatchSource:0}: Error finding container 81f31016c1cba082b07d34c2d65a70faf6f6b3779acaefcaaf756aa2037d04fb: Status 404 returned error can't find the container with id 81f31016c1cba082b07d34c2d65a70faf6f6b3779acaefcaaf756aa2037d04fb Dec 09 17:40:45 crc kubenswrapper[4954]: I1209 17:40:45.466425 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg6qk" event={"ID":"4d9f331b-1de4-4d06-8caa-7115a9e1c10e","Type":"ContainerStarted","Data":"81f31016c1cba082b07d34c2d65a70faf6f6b3779acaefcaaf756aa2037d04fb"} Dec 09 17:40:45 crc kubenswrapper[4954]: W1209 17:40:45.519144 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda20b6b58_cb72_432d_8ec8_7c7f93265d36.slice/crio-307d4f5d263167ff5f0bdd97e170cfc11dd53b0e80fc679c2fc12fd7274f9385 WatchSource:0}: Error finding container 307d4f5d263167ff5f0bdd97e170cfc11dd53b0e80fc679c2fc12fd7274f9385: Status 404 returned error can't find the container with id 307d4f5d263167ff5f0bdd97e170cfc11dd53b0e80fc679c2fc12fd7274f9385 Dec 09 17:40:45 crc kubenswrapper[4954]: I1209 17:40:45.521837 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl"] Dec 09 17:40:46 crc kubenswrapper[4954]: I1209 17:40:46.476808 4954 generic.go:334] "Generic (PLEG): container finished" podID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerID="61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c" exitCode=0 Dec 09 17:40:46 crc kubenswrapper[4954]: I1209 17:40:46.476893 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg6qk" event={"ID":"4d9f331b-1de4-4d06-8caa-7115a9e1c10e","Type":"ContainerDied","Data":"61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c"} Dec 09 17:40:46 crc kubenswrapper[4954]: I1209 17:40:46.479242 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" event={"ID":"a20b6b58-cb72-432d-8ec8-7c7f93265d36","Type":"ContainerStarted","Data":"307d4f5d263167ff5f0bdd97e170cfc11dd53b0e80fc679c2fc12fd7274f9385"} Dec 09 17:40:47 crc kubenswrapper[4954]: I1209 17:40:47.121277 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:40:47 crc kubenswrapper[4954]: E1209 17:40:47.121628 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:40:47 crc kubenswrapper[4954]: I1209 17:40:47.498686 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" event={"ID":"a20b6b58-cb72-432d-8ec8-7c7f93265d36","Type":"ContainerStarted","Data":"efa0128709049203031039bacd70b5d5817045c7b0b786436adc30e2ff74e56b"} Dec 09 17:40:47 crc kubenswrapper[4954]: I1209 17:40:47.531550 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" podStartSLOduration=2.745430136 podStartE2EDuration="3.531528781s" podCreationTimestamp="2025-12-09 17:40:44 +0000 UTC" firstStartedPulling="2025-12-09 17:40:45.522696889 +0000 UTC m=+2641.910870709" lastFinishedPulling="2025-12-09 17:40:46.308795534 +0000 UTC m=+2642.696969354" observedRunningTime="2025-12-09 17:40:47.519820593 +0000 UTC m=+2643.907994413" watchObservedRunningTime="2025-12-09 17:40:47.531528781 +0000 UTC m=+2643.919702601" Dec 09 17:40:48 crc kubenswrapper[4954]: I1209 17:40:48.509139 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg6qk" event={"ID":"4d9f331b-1de4-4d06-8caa-7115a9e1c10e","Type":"ContainerStarted","Data":"b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5"} Dec 09 17:40:49 crc kubenswrapper[4954]: I1209 17:40:49.527877 4954 generic.go:334] "Generic (PLEG): container finished" podID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerID="b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5" exitCode=0 Dec 09 17:40:49 crc kubenswrapper[4954]: I1209 17:40:49.527947 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg6qk" event={"ID":"4d9f331b-1de4-4d06-8caa-7115a9e1c10e","Type":"ContainerDied","Data":"b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5"} Dec 09 17:40:50 crc kubenswrapper[4954]: I1209 17:40:50.545649 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg6qk" event={"ID":"4d9f331b-1de4-4d06-8caa-7115a9e1c10e","Type":"ContainerStarted","Data":"db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12"} Dec 09 17:40:50 crc kubenswrapper[4954]: I1209 17:40:50.565246 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xg6qk" podStartSLOduration=2.750556518 podStartE2EDuration="6.565228967s" podCreationTimestamp="2025-12-09 17:40:44 +0000 UTC" firstStartedPulling="2025-12-09 17:40:46.479377599 +0000 UTC m=+2642.867551419" lastFinishedPulling="2025-12-09 17:40:50.294050048 +0000 UTC m=+2646.682223868" observedRunningTime="2025-12-09 17:40:50.56180249 +0000 UTC m=+2646.949976330" watchObservedRunningTime="2025-12-09 17:40:50.565228967 +0000 UTC m=+2646.953402787" Dec 09 17:40:54 crc kubenswrapper[4954]: I1209 17:40:54.829631 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:54 crc kubenswrapper[4954]: I1209 17:40:54.830987 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:54 crc kubenswrapper[4954]: I1209 17:40:54.878337 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:55 crc kubenswrapper[4954]: E1209 17:40:55.122615 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:40:55 crc kubenswrapper[4954]: I1209 17:40:55.636002 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:56 crc kubenswrapper[4954]: I1209 17:40:56.445503 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xg6qk"] Dec 09 17:40:57 crc kubenswrapper[4954]: I1209 17:40:57.608712 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xg6qk" podUID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerName="registry-server" containerID="cri-o://db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12" gracePeriod=2 Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.087497 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:58 crc kubenswrapper[4954]: E1209 17:40:58.122036 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.132534 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nthjj\" (UniqueName: \"kubernetes.io/projected/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-kube-api-access-nthjj\") pod \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.132712 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-utilities\") pod \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.132793 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-catalog-content\") pod \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\" (UID: \"4d9f331b-1de4-4d06-8caa-7115a9e1c10e\") " Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.134815 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-utilities" (OuterVolumeSpecName: "utilities") pod "4d9f331b-1de4-4d06-8caa-7115a9e1c10e" (UID: "4d9f331b-1de4-4d06-8caa-7115a9e1c10e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.140776 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-kube-api-access-nthjj" (OuterVolumeSpecName: "kube-api-access-nthjj") pod "4d9f331b-1de4-4d06-8caa-7115a9e1c10e" (UID: "4d9f331b-1de4-4d06-8caa-7115a9e1c10e"). InnerVolumeSpecName "kube-api-access-nthjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.193807 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d9f331b-1de4-4d06-8caa-7115a9e1c10e" (UID: "4d9f331b-1de4-4d06-8caa-7115a9e1c10e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.236082 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nthjj\" (UniqueName: \"kubernetes.io/projected/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-kube-api-access-nthjj\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.236111 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.236121 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d9f331b-1de4-4d06-8caa-7115a9e1c10e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.619868 4954 generic.go:334] "Generic (PLEG): container finished" podID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerID="db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12" exitCode=0 Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.619919 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xg6qk" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.619919 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg6qk" event={"ID":"4d9f331b-1de4-4d06-8caa-7115a9e1c10e","Type":"ContainerDied","Data":"db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12"} Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.620004 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xg6qk" event={"ID":"4d9f331b-1de4-4d06-8caa-7115a9e1c10e","Type":"ContainerDied","Data":"81f31016c1cba082b07d34c2d65a70faf6f6b3779acaefcaaf756aa2037d04fb"} Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.620024 4954 scope.go:117] "RemoveContainer" containerID="db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.655700 4954 scope.go:117] "RemoveContainer" containerID="b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.664559 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xg6qk"] Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.672588 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xg6qk"] Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.676541 4954 scope.go:117] "RemoveContainer" containerID="61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.737302 4954 scope.go:117] "RemoveContainer" containerID="db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12" Dec 09 17:40:58 crc kubenswrapper[4954]: E1209 17:40:58.737961 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12\": container with ID starting with db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12 not found: ID does not exist" containerID="db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.737998 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12"} err="failed to get container status \"db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12\": rpc error: code = NotFound desc = could not find container \"db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12\": container with ID starting with db2cab29dde04df6f152597f9602823ef605a24cf0538572329fc294f2f69c12 not found: ID does not exist" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.738039 4954 scope.go:117] "RemoveContainer" containerID="b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5" Dec 09 17:40:58 crc kubenswrapper[4954]: E1209 17:40:58.738373 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5\": container with ID starting with b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5 not found: ID does not exist" containerID="b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.738399 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5"} err="failed to get container status \"b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5\": rpc error: code = NotFound desc = could not find container \"b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5\": container with ID starting with b0e1d348c228887bcafe3391c8fccca4cce5c4afa0a46deff9f44920199172c5 not found: ID does not exist" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.738420 4954 scope.go:117] "RemoveContainer" containerID="61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c" Dec 09 17:40:58 crc kubenswrapper[4954]: E1209 17:40:58.738743 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c\": container with ID starting with 61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c not found: ID does not exist" containerID="61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c" Dec 09 17:40:58 crc kubenswrapper[4954]: I1209 17:40:58.738783 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c"} err="failed to get container status \"61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c\": rpc error: code = NotFound desc = could not find container \"61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c\": container with ID starting with 61f86879e2370ac08bce8b972b3ccb4c23dd138174b6c37b208110f672b8ec1c not found: ID does not exist" Dec 09 17:41:00 crc kubenswrapper[4954]: I1209 17:41:00.135811 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" path="/var/lib/kubelet/pods/4d9f331b-1de4-4d06-8caa-7115a9e1c10e/volumes" Dec 09 17:41:01 crc kubenswrapper[4954]: I1209 17:41:01.120511 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:41:01 crc kubenswrapper[4954]: E1209 17:41:01.120857 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:41:04 crc kubenswrapper[4954]: I1209 17:41:04.955396 4954 scope.go:117] "RemoveContainer" containerID="1060380ad794329e8565edb3dda2d6962aeb50e7069307bca0a9fc23e956bb26" Dec 09 17:41:06 crc kubenswrapper[4954]: E1209 17:41:06.123353 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:41:10 crc kubenswrapper[4954]: E1209 17:41:10.122710 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:41:15 crc kubenswrapper[4954]: I1209 17:41:15.120654 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:41:15 crc kubenswrapper[4954]: E1209 17:41:15.121386 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:41:18 crc kubenswrapper[4954]: E1209 17:41:18.127202 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:41:25 crc kubenswrapper[4954]: E1209 17:41:25.123394 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:41:29 crc kubenswrapper[4954]: E1209 17:41:29.124679 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:41:30 crc kubenswrapper[4954]: I1209 17:41:30.122373 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:41:30 crc kubenswrapper[4954]: E1209 17:41:30.122705 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:41:37 crc kubenswrapper[4954]: E1209 17:41:37.122198 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:41:41 crc kubenswrapper[4954]: E1209 17:41:41.124057 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:41:45 crc kubenswrapper[4954]: I1209 17:41:45.120535 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:41:46 crc kubenswrapper[4954]: I1209 17:41:46.053442 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"8b51db8880937853b1d1e5ce56a6c02a6aa8e1062a55e02266d9eb62deba3e8f"} Dec 09 17:41:49 crc kubenswrapper[4954]: E1209 17:41:49.124699 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:41:54 crc kubenswrapper[4954]: E1209 17:41:54.122092 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:41:56 crc kubenswrapper[4954]: I1209 17:41:56.145970 4954 generic.go:334] "Generic (PLEG): container finished" podID="a20b6b58-cb72-432d-8ec8-7c7f93265d36" containerID="efa0128709049203031039bacd70b5d5817045c7b0b786436adc30e2ff74e56b" exitCode=0 Dec 09 17:41:56 crc kubenswrapper[4954]: I1209 17:41:56.146046 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" event={"ID":"a20b6b58-cb72-432d-8ec8-7c7f93265d36","Type":"ContainerDied","Data":"efa0128709049203031039bacd70b5d5817045c7b0b786436adc30e2ff74e56b"} Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.594653 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.761175 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-inventory\") pod \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.761495 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ssh-key\") pod \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.761618 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovn-combined-ca-bundle\") pod \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.761641 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pzn9\" (UniqueName: \"kubernetes.io/projected/a20b6b58-cb72-432d-8ec8-7c7f93265d36-kube-api-access-2pzn9\") pod \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.761661 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovncontroller-config-0\") pod \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\" (UID: \"a20b6b58-cb72-432d-8ec8-7c7f93265d36\") " Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.766536 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a20b6b58-cb72-432d-8ec8-7c7f93265d36" (UID: "a20b6b58-cb72-432d-8ec8-7c7f93265d36"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.767138 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a20b6b58-cb72-432d-8ec8-7c7f93265d36-kube-api-access-2pzn9" (OuterVolumeSpecName: "kube-api-access-2pzn9") pod "a20b6b58-cb72-432d-8ec8-7c7f93265d36" (UID: "a20b6b58-cb72-432d-8ec8-7c7f93265d36"). InnerVolumeSpecName "kube-api-access-2pzn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.788447 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "a20b6b58-cb72-432d-8ec8-7c7f93265d36" (UID: "a20b6b58-cb72-432d-8ec8-7c7f93265d36"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.790586 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a20b6b58-cb72-432d-8ec8-7c7f93265d36" (UID: "a20b6b58-cb72-432d-8ec8-7c7f93265d36"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.794128 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-inventory" (OuterVolumeSpecName: "inventory") pod "a20b6b58-cb72-432d-8ec8-7c7f93265d36" (UID: "a20b6b58-cb72-432d-8ec8-7c7f93265d36"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.864654 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.864698 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.864712 4954 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.864727 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pzn9\" (UniqueName: \"kubernetes.io/projected/a20b6b58-cb72-432d-8ec8-7c7f93265d36-kube-api-access-2pzn9\") on node \"crc\" DevicePath \"\"" Dec 09 17:41:57 crc kubenswrapper[4954]: I1209 17:41:57.864736 4954 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/a20b6b58-cb72-432d-8ec8-7c7f93265d36-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.170530 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" event={"ID":"a20b6b58-cb72-432d-8ec8-7c7f93265d36","Type":"ContainerDied","Data":"307d4f5d263167ff5f0bdd97e170cfc11dd53b0e80fc679c2fc12fd7274f9385"} Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.170579 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="307d4f5d263167ff5f0bdd97e170cfc11dd53b0e80fc679c2fc12fd7274f9385" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.170646 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jt7gl" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.251991 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k"] Dec 09 17:41:58 crc kubenswrapper[4954]: E1209 17:41:58.252509 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerName="extract-content" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.252526 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerName="extract-content" Dec 09 17:41:58 crc kubenswrapper[4954]: E1209 17:41:58.252547 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerName="registry-server" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.252554 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerName="registry-server" Dec 09 17:41:58 crc kubenswrapper[4954]: E1209 17:41:58.252569 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerName="extract-utilities" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.252576 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerName="extract-utilities" Dec 09 17:41:58 crc kubenswrapper[4954]: E1209 17:41:58.252669 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a20b6b58-cb72-432d-8ec8-7c7f93265d36" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.252677 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a20b6b58-cb72-432d-8ec8-7c7f93265d36" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.252870 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a20b6b58-cb72-432d-8ec8-7c7f93265d36" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.252893 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9f331b-1de4-4d06-8caa-7115a9e1c10e" containerName="registry-server" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.253691 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.256032 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.256350 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.256471 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.257251 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.259116 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.262470 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k"] Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.374424 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.374477 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.374555 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.374654 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.374775 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dj8v\" (UniqueName: \"kubernetes.io/projected/d64c4820-0dfe-46fa-9275-e9d45bbb268e-kube-api-access-8dj8v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.476664 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dj8v\" (UniqueName: \"kubernetes.io/projected/d64c4820-0dfe-46fa-9275-e9d45bbb268e-kube-api-access-8dj8v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.476734 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.476764 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.476822 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.476904 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.481696 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.482139 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.482477 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.485862 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.494880 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dj8v\" (UniqueName: \"kubernetes.io/projected/d64c4820-0dfe-46fa-9275-e9d45bbb268e-kube-api-access-8dj8v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:58 crc kubenswrapper[4954]: I1209 17:41:58.572573 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:41:59 crc kubenswrapper[4954]: I1209 17:41:59.127122 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k"] Dec 09 17:41:59 crc kubenswrapper[4954]: I1209 17:41:59.130545 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 17:41:59 crc kubenswrapper[4954]: I1209 17:41:59.181899 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" event={"ID":"d64c4820-0dfe-46fa-9275-e9d45bbb268e","Type":"ContainerStarted","Data":"1459d24c659da2e0650584d10ecf0dc12c4645620e41e9deae4cc3d10efb3062"} Dec 09 17:42:01 crc kubenswrapper[4954]: I1209 17:42:01.205234 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" event={"ID":"d64c4820-0dfe-46fa-9275-e9d45bbb268e","Type":"ContainerStarted","Data":"06bebe0dc981564d98ab2770c730bfb639e65078b279977d414f77e4c7422aff"} Dec 09 17:42:01 crc kubenswrapper[4954]: I1209 17:42:01.233667 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" podStartSLOduration=2.251518544 podStartE2EDuration="3.233634553s" podCreationTimestamp="2025-12-09 17:41:58 +0000 UTC" firstStartedPulling="2025-12-09 17:41:59.130308874 +0000 UTC m=+2715.518482694" lastFinishedPulling="2025-12-09 17:42:00.112424873 +0000 UTC m=+2716.500598703" observedRunningTime="2025-12-09 17:42:01.222951677 +0000 UTC m=+2717.611125497" watchObservedRunningTime="2025-12-09 17:42:01.233634553 +0000 UTC m=+2717.621808373" Dec 09 17:42:03 crc kubenswrapper[4954]: E1209 17:42:03.122232 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:42:07 crc kubenswrapper[4954]: E1209 17:42:07.121934 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:42:18 crc kubenswrapper[4954]: E1209 17:42:18.124079 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:42:21 crc kubenswrapper[4954]: E1209 17:42:21.123173 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:42:30 crc kubenswrapper[4954]: E1209 17:42:30.252531 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:42:30 crc kubenswrapper[4954]: E1209 17:42:30.253107 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:42:30 crc kubenswrapper[4954]: E1209 17:42:30.253244 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:42:30 crc kubenswrapper[4954]: E1209 17:42:30.254406 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:42:34 crc kubenswrapper[4954]: E1209 17:42:34.220880 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:42:34 crc kubenswrapper[4954]: E1209 17:42:34.221455 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:42:34 crc kubenswrapper[4954]: E1209 17:42:34.221588 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:42:34 crc kubenswrapper[4954]: E1209 17:42:34.223077 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:42:41 crc kubenswrapper[4954]: E1209 17:42:41.123110 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:42:48 crc kubenswrapper[4954]: E1209 17:42:48.125089 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:42:54 crc kubenswrapper[4954]: E1209 17:42:54.131386 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:43:03 crc kubenswrapper[4954]: E1209 17:43:03.123303 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.020816 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h5868"] Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.027366 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.040692 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5868"] Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.183832 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-utilities\") pod \"redhat-marketplace-h5868\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.184042 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-catalog-content\") pod \"redhat-marketplace-h5868\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.184137 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hctjv\" (UniqueName: \"kubernetes.io/projected/2806b077-5dd4-4bc1-97fe-563da754aed2-kube-api-access-hctjv\") pod \"redhat-marketplace-h5868\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.285630 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-catalog-content\") pod \"redhat-marketplace-h5868\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.286042 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hctjv\" (UniqueName: \"kubernetes.io/projected/2806b077-5dd4-4bc1-97fe-563da754aed2-kube-api-access-hctjv\") pod \"redhat-marketplace-h5868\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.286111 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-utilities\") pod \"redhat-marketplace-h5868\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.286209 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-catalog-content\") pod \"redhat-marketplace-h5868\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.286573 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-utilities\") pod \"redhat-marketplace-h5868\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.343404 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hctjv\" (UniqueName: \"kubernetes.io/projected/2806b077-5dd4-4bc1-97fe-563da754aed2-kube-api-access-hctjv\") pod \"redhat-marketplace-h5868\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.351321 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:05 crc kubenswrapper[4954]: I1209 17:43:05.898936 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5868"] Dec 09 17:43:06 crc kubenswrapper[4954]: I1209 17:43:06.838483 4954 generic.go:334] "Generic (PLEG): container finished" podID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerID="ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331" exitCode=0 Dec 09 17:43:06 crc kubenswrapper[4954]: I1209 17:43:06.838714 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5868" event={"ID":"2806b077-5dd4-4bc1-97fe-563da754aed2","Type":"ContainerDied","Data":"ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331"} Dec 09 17:43:06 crc kubenswrapper[4954]: I1209 17:43:06.838829 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5868" event={"ID":"2806b077-5dd4-4bc1-97fe-563da754aed2","Type":"ContainerStarted","Data":"87413e35c2b551f820c30b1e5531b4baa16740d19f9453cb4d3af758ca99f021"} Dec 09 17:43:07 crc kubenswrapper[4954]: E1209 17:43:07.122328 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:43:07 crc kubenswrapper[4954]: I1209 17:43:07.849405 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5868" event={"ID":"2806b077-5dd4-4bc1-97fe-563da754aed2","Type":"ContainerStarted","Data":"e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d"} Dec 09 17:43:08 crc kubenswrapper[4954]: I1209 17:43:08.861978 4954 generic.go:334] "Generic (PLEG): container finished" podID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerID="e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d" exitCode=0 Dec 09 17:43:08 crc kubenswrapper[4954]: I1209 17:43:08.862058 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5868" event={"ID":"2806b077-5dd4-4bc1-97fe-563da754aed2","Type":"ContainerDied","Data":"e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d"} Dec 09 17:43:10 crc kubenswrapper[4954]: I1209 17:43:10.882414 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5868" event={"ID":"2806b077-5dd4-4bc1-97fe-563da754aed2","Type":"ContainerStarted","Data":"610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d"} Dec 09 17:43:10 crc kubenswrapper[4954]: I1209 17:43:10.905268 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h5868" podStartSLOduration=3.79961567 podStartE2EDuration="6.905248364s" podCreationTimestamp="2025-12-09 17:43:04 +0000 UTC" firstStartedPulling="2025-12-09 17:43:06.840893918 +0000 UTC m=+2783.229067738" lastFinishedPulling="2025-12-09 17:43:09.946526612 +0000 UTC m=+2786.334700432" observedRunningTime="2025-12-09 17:43:10.899913106 +0000 UTC m=+2787.288086936" watchObservedRunningTime="2025-12-09 17:43:10.905248364 +0000 UTC m=+2787.293422184" Dec 09 17:43:15 crc kubenswrapper[4954]: I1209 17:43:15.351727 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:15 crc kubenswrapper[4954]: I1209 17:43:15.352283 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:15 crc kubenswrapper[4954]: I1209 17:43:15.487042 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:15 crc kubenswrapper[4954]: I1209 17:43:15.983303 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:16 crc kubenswrapper[4954]: I1209 17:43:16.042104 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5868"] Dec 09 17:43:17 crc kubenswrapper[4954]: I1209 17:43:17.943986 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h5868" podUID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerName="registry-server" containerID="cri-o://610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d" gracePeriod=2 Dec 09 17:43:18 crc kubenswrapper[4954]: E1209 17:43:18.122443 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:43:18 crc kubenswrapper[4954]: E1209 17:43:18.122741 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.461725 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.611572 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hctjv\" (UniqueName: \"kubernetes.io/projected/2806b077-5dd4-4bc1-97fe-563da754aed2-kube-api-access-hctjv\") pod \"2806b077-5dd4-4bc1-97fe-563da754aed2\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.611813 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-catalog-content\") pod \"2806b077-5dd4-4bc1-97fe-563da754aed2\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.611894 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-utilities\") pod \"2806b077-5dd4-4bc1-97fe-563da754aed2\" (UID: \"2806b077-5dd4-4bc1-97fe-563da754aed2\") " Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.612952 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-utilities" (OuterVolumeSpecName: "utilities") pod "2806b077-5dd4-4bc1-97fe-563da754aed2" (UID: "2806b077-5dd4-4bc1-97fe-563da754aed2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.620786 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2806b077-5dd4-4bc1-97fe-563da754aed2-kube-api-access-hctjv" (OuterVolumeSpecName: "kube-api-access-hctjv") pod "2806b077-5dd4-4bc1-97fe-563da754aed2" (UID: "2806b077-5dd4-4bc1-97fe-563da754aed2"). InnerVolumeSpecName "kube-api-access-hctjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.631343 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2806b077-5dd4-4bc1-97fe-563da754aed2" (UID: "2806b077-5dd4-4bc1-97fe-563da754aed2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.715143 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hctjv\" (UniqueName: \"kubernetes.io/projected/2806b077-5dd4-4bc1-97fe-563da754aed2-kube-api-access-hctjv\") on node \"crc\" DevicePath \"\"" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.715188 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.715203 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2806b077-5dd4-4bc1-97fe-563da754aed2-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.956862 4954 generic.go:334] "Generic (PLEG): container finished" podID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerID="610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d" exitCode=0 Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.956940 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5868" event={"ID":"2806b077-5dd4-4bc1-97fe-563da754aed2","Type":"ContainerDied","Data":"610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d"} Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.957008 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5868" event={"ID":"2806b077-5dd4-4bc1-97fe-563da754aed2","Type":"ContainerDied","Data":"87413e35c2b551f820c30b1e5531b4baa16740d19f9453cb4d3af758ca99f021"} Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.957018 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5868" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.957032 4954 scope.go:117] "RemoveContainer" containerID="610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.981835 4954 scope.go:117] "RemoveContainer" containerID="e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d" Dec 09 17:43:18 crc kubenswrapper[4954]: I1209 17:43:18.995765 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5868"] Dec 09 17:43:19 crc kubenswrapper[4954]: I1209 17:43:19.004426 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5868"] Dec 09 17:43:19 crc kubenswrapper[4954]: I1209 17:43:19.021826 4954 scope.go:117] "RemoveContainer" containerID="ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331" Dec 09 17:43:19 crc kubenswrapper[4954]: I1209 17:43:19.062456 4954 scope.go:117] "RemoveContainer" containerID="610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d" Dec 09 17:43:19 crc kubenswrapper[4954]: E1209 17:43:19.063582 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d\": container with ID starting with 610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d not found: ID does not exist" containerID="610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d" Dec 09 17:43:19 crc kubenswrapper[4954]: I1209 17:43:19.063639 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d"} err="failed to get container status \"610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d\": rpc error: code = NotFound desc = could not find container \"610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d\": container with ID starting with 610df063db2a12804aa838cd6c06e2ab8856a1ce4659311c2d0f0ecdcc9ce21d not found: ID does not exist" Dec 09 17:43:19 crc kubenswrapper[4954]: I1209 17:43:19.063668 4954 scope.go:117] "RemoveContainer" containerID="e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d" Dec 09 17:43:19 crc kubenswrapper[4954]: E1209 17:43:19.064107 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d\": container with ID starting with e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d not found: ID does not exist" containerID="e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d" Dec 09 17:43:19 crc kubenswrapper[4954]: I1209 17:43:19.064142 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d"} err="failed to get container status \"e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d\": rpc error: code = NotFound desc = could not find container \"e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d\": container with ID starting with e7bb851115156a1613d27cf5042b6776eb72ebf51f6057c986b6145a434bd38d not found: ID does not exist" Dec 09 17:43:19 crc kubenswrapper[4954]: I1209 17:43:19.064162 4954 scope.go:117] "RemoveContainer" containerID="ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331" Dec 09 17:43:19 crc kubenswrapper[4954]: E1209 17:43:19.064868 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331\": container with ID starting with ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331 not found: ID does not exist" containerID="ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331" Dec 09 17:43:19 crc kubenswrapper[4954]: I1209 17:43:19.064892 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331"} err="failed to get container status \"ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331\": rpc error: code = NotFound desc = could not find container \"ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331\": container with ID starting with ad9512c2d1aa930c85ce90abb06b543f1867f5b66cab20c848b862da49fab331 not found: ID does not exist" Dec 09 17:43:20 crc kubenswrapper[4954]: I1209 17:43:20.156617 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2806b077-5dd4-4bc1-97fe-563da754aed2" path="/var/lib/kubelet/pods/2806b077-5dd4-4bc1-97fe-563da754aed2/volumes" Dec 09 17:43:30 crc kubenswrapper[4954]: E1209 17:43:30.123924 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:43:32 crc kubenswrapper[4954]: E1209 17:43:32.123284 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:43:45 crc kubenswrapper[4954]: E1209 17:43:45.124187 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:43:47 crc kubenswrapper[4954]: E1209 17:43:47.122441 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:43:57 crc kubenswrapper[4954]: E1209 17:43:57.122775 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:43:58 crc kubenswrapper[4954]: E1209 17:43:58.126497 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.319605 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pvczg"] Dec 09 17:44:02 crc kubenswrapper[4954]: E1209 17:44:02.320496 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerName="extract-utilities" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.320518 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerName="extract-utilities" Dec 09 17:44:02 crc kubenswrapper[4954]: E1209 17:44:02.320536 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerName="extract-content" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.320545 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerName="extract-content" Dec 09 17:44:02 crc kubenswrapper[4954]: E1209 17:44:02.320685 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerName="registry-server" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.320698 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerName="registry-server" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.320991 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="2806b077-5dd4-4bc1-97fe-563da754aed2" containerName="registry-server" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.323052 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.331875 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pvczg"] Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.426896 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-utilities\") pod \"certified-operators-pvczg\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.427232 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hn5d\" (UniqueName: \"kubernetes.io/projected/84354081-af02-4bbf-8ad1-9d2cf8ae2621-kube-api-access-4hn5d\") pod \"certified-operators-pvczg\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.427467 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-catalog-content\") pod \"certified-operators-pvczg\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.529515 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-utilities\") pod \"certified-operators-pvczg\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.529793 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hn5d\" (UniqueName: \"kubernetes.io/projected/84354081-af02-4bbf-8ad1-9d2cf8ae2621-kube-api-access-4hn5d\") pod \"certified-operators-pvczg\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.529877 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-catalog-content\") pod \"certified-operators-pvczg\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.530058 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-utilities\") pod \"certified-operators-pvczg\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.530333 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-catalog-content\") pod \"certified-operators-pvczg\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.553549 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hn5d\" (UniqueName: \"kubernetes.io/projected/84354081-af02-4bbf-8ad1-9d2cf8ae2621-kube-api-access-4hn5d\") pod \"certified-operators-pvczg\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:02 crc kubenswrapper[4954]: I1209 17:44:02.646091 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:03 crc kubenswrapper[4954]: I1209 17:44:03.191165 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pvczg"] Dec 09 17:44:03 crc kubenswrapper[4954]: I1209 17:44:03.402498 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvczg" event={"ID":"84354081-af02-4bbf-8ad1-9d2cf8ae2621","Type":"ContainerStarted","Data":"b7541fca74b5c21fdadda0a2146b5a551e3781b074052bd5f172d75f95242323"} Dec 09 17:44:04 crc kubenswrapper[4954]: I1209 17:44:04.413120 4954 generic.go:334] "Generic (PLEG): container finished" podID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerID="ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60" exitCode=0 Dec 09 17:44:04 crc kubenswrapper[4954]: I1209 17:44:04.413171 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvczg" event={"ID":"84354081-af02-4bbf-8ad1-9d2cf8ae2621","Type":"ContainerDied","Data":"ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60"} Dec 09 17:44:06 crc kubenswrapper[4954]: I1209 17:44:06.433096 4954 generic.go:334] "Generic (PLEG): container finished" podID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerID="32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae" exitCode=0 Dec 09 17:44:06 crc kubenswrapper[4954]: I1209 17:44:06.433203 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvczg" event={"ID":"84354081-af02-4bbf-8ad1-9d2cf8ae2621","Type":"ContainerDied","Data":"32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae"} Dec 09 17:44:08 crc kubenswrapper[4954]: I1209 17:44:08.451460 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvczg" event={"ID":"84354081-af02-4bbf-8ad1-9d2cf8ae2621","Type":"ContainerStarted","Data":"d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6"} Dec 09 17:44:08 crc kubenswrapper[4954]: I1209 17:44:08.473571 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pvczg" podStartSLOduration=2.789801946 podStartE2EDuration="6.473550401s" podCreationTimestamp="2025-12-09 17:44:02 +0000 UTC" firstStartedPulling="2025-12-09 17:44:04.415327579 +0000 UTC m=+2840.803501399" lastFinishedPulling="2025-12-09 17:44:08.099076034 +0000 UTC m=+2844.487249854" observedRunningTime="2025-12-09 17:44:08.469341188 +0000 UTC m=+2844.857515038" watchObservedRunningTime="2025-12-09 17:44:08.473550401 +0000 UTC m=+2844.861724221" Dec 09 17:44:09 crc kubenswrapper[4954]: E1209 17:44:09.121732 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:44:12 crc kubenswrapper[4954]: E1209 17:44:12.122482 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:44:12 crc kubenswrapper[4954]: I1209 17:44:12.646903 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:12 crc kubenswrapper[4954]: I1209 17:44:12.646975 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:12 crc kubenswrapper[4954]: I1209 17:44:12.694473 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.307778 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xcqt7"] Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.310891 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.318303 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xcqt7"] Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.480534 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-utilities\") pod \"redhat-operators-xcqt7\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.480687 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-catalog-content\") pod \"redhat-operators-xcqt7\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.480721 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t8hp\" (UniqueName: \"kubernetes.io/projected/26115b90-b06a-4f5c-8586-6d4cbe3c2869-kube-api-access-9t8hp\") pod \"redhat-operators-xcqt7\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.551904 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.583052 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-utilities\") pod \"redhat-operators-xcqt7\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.583214 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-catalog-content\") pod \"redhat-operators-xcqt7\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.583243 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t8hp\" (UniqueName: \"kubernetes.io/projected/26115b90-b06a-4f5c-8586-6d4cbe3c2869-kube-api-access-9t8hp\") pod \"redhat-operators-xcqt7\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.583573 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-utilities\") pod \"redhat-operators-xcqt7\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.583850 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-catalog-content\") pod \"redhat-operators-xcqt7\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.607471 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t8hp\" (UniqueName: \"kubernetes.io/projected/26115b90-b06a-4f5c-8586-6d4cbe3c2869-kube-api-access-9t8hp\") pod \"redhat-operators-xcqt7\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.644675 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.756734 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:44:13 crc kubenswrapper[4954]: I1209 17:44:13.757026 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:44:14 crc kubenswrapper[4954]: I1209 17:44:14.148670 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xcqt7"] Dec 09 17:44:14 crc kubenswrapper[4954]: I1209 17:44:14.513449 4954 generic.go:334] "Generic (PLEG): container finished" podID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerID="90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862" exitCode=0 Dec 09 17:44:14 crc kubenswrapper[4954]: I1209 17:44:14.513562 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcqt7" event={"ID":"26115b90-b06a-4f5c-8586-6d4cbe3c2869","Type":"ContainerDied","Data":"90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862"} Dec 09 17:44:14 crc kubenswrapper[4954]: I1209 17:44:14.513627 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcqt7" event={"ID":"26115b90-b06a-4f5c-8586-6d4cbe3c2869","Type":"ContainerStarted","Data":"3fd263d89e23b30e16159d8d9c6b2b7b0478a187f23e87281aa4b9a44111cf80"} Dec 09 17:44:15 crc kubenswrapper[4954]: I1209 17:44:15.525893 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcqt7" event={"ID":"26115b90-b06a-4f5c-8586-6d4cbe3c2869","Type":"ContainerStarted","Data":"6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3"} Dec 09 17:44:15 crc kubenswrapper[4954]: I1209 17:44:15.710445 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pvczg"] Dec 09 17:44:15 crc kubenswrapper[4954]: I1209 17:44:15.711003 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pvczg" podUID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerName="registry-server" containerID="cri-o://d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6" gracePeriod=2 Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.205730 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.349101 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-catalog-content\") pod \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.349443 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-utilities\") pod \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.349512 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hn5d\" (UniqueName: \"kubernetes.io/projected/84354081-af02-4bbf-8ad1-9d2cf8ae2621-kube-api-access-4hn5d\") pod \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\" (UID: \"84354081-af02-4bbf-8ad1-9d2cf8ae2621\") " Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.351797 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-utilities" (OuterVolumeSpecName: "utilities") pod "84354081-af02-4bbf-8ad1-9d2cf8ae2621" (UID: "84354081-af02-4bbf-8ad1-9d2cf8ae2621"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.356128 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84354081-af02-4bbf-8ad1-9d2cf8ae2621-kube-api-access-4hn5d" (OuterVolumeSpecName: "kube-api-access-4hn5d") pod "84354081-af02-4bbf-8ad1-9d2cf8ae2621" (UID: "84354081-af02-4bbf-8ad1-9d2cf8ae2621"). InnerVolumeSpecName "kube-api-access-4hn5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.394715 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84354081-af02-4bbf-8ad1-9d2cf8ae2621" (UID: "84354081-af02-4bbf-8ad1-9d2cf8ae2621"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.451965 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.452001 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84354081-af02-4bbf-8ad1-9d2cf8ae2621-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.452011 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hn5d\" (UniqueName: \"kubernetes.io/projected/84354081-af02-4bbf-8ad1-9d2cf8ae2621-kube-api-access-4hn5d\") on node \"crc\" DevicePath \"\"" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.543682 4954 generic.go:334] "Generic (PLEG): container finished" podID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerID="d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6" exitCode=0 Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.544131 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvczg" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.544295 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvczg" event={"ID":"84354081-af02-4bbf-8ad1-9d2cf8ae2621","Type":"ContainerDied","Data":"d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6"} Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.544351 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvczg" event={"ID":"84354081-af02-4bbf-8ad1-9d2cf8ae2621","Type":"ContainerDied","Data":"b7541fca74b5c21fdadda0a2146b5a551e3781b074052bd5f172d75f95242323"} Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.544373 4954 scope.go:117] "RemoveContainer" containerID="d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.586431 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pvczg"] Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.591978 4954 scope.go:117] "RemoveContainer" containerID="32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.599092 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pvczg"] Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.617840 4954 scope.go:117] "RemoveContainer" containerID="ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.661360 4954 scope.go:117] "RemoveContainer" containerID="d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6" Dec 09 17:44:16 crc kubenswrapper[4954]: E1209 17:44:16.661793 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6\": container with ID starting with d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6 not found: ID does not exist" containerID="d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.661823 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6"} err="failed to get container status \"d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6\": rpc error: code = NotFound desc = could not find container \"d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6\": container with ID starting with d0665d15e282e04308f26152ccae34c7ba40855f1e0c0473c04502f427d307b6 not found: ID does not exist" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.661846 4954 scope.go:117] "RemoveContainer" containerID="32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae" Dec 09 17:44:16 crc kubenswrapper[4954]: E1209 17:44:16.662150 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae\": container with ID starting with 32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae not found: ID does not exist" containerID="32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.662182 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae"} err="failed to get container status \"32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae\": rpc error: code = NotFound desc = could not find container \"32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae\": container with ID starting with 32ee3bb613c32485e67ac7fa7c9f565298932cbdd39eb9af767bb43e9832f8ae not found: ID does not exist" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.662202 4954 scope.go:117] "RemoveContainer" containerID="ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60" Dec 09 17:44:16 crc kubenswrapper[4954]: E1209 17:44:16.662499 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60\": container with ID starting with ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60 not found: ID does not exist" containerID="ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60" Dec 09 17:44:16 crc kubenswrapper[4954]: I1209 17:44:16.662549 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60"} err="failed to get container status \"ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60\": rpc error: code = NotFound desc = could not find container \"ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60\": container with ID starting with ade25c3a736ad489dfc13a6e5bbe7a09b164de09385cbbae0c224ed3caed0f60 not found: ID does not exist" Dec 09 17:44:18 crc kubenswrapper[4954]: I1209 17:44:18.136118 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" path="/var/lib/kubelet/pods/84354081-af02-4bbf-8ad1-9d2cf8ae2621/volumes" Dec 09 17:44:18 crc kubenswrapper[4954]: I1209 17:44:18.566639 4954 generic.go:334] "Generic (PLEG): container finished" podID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerID="6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3" exitCode=0 Dec 09 17:44:18 crc kubenswrapper[4954]: I1209 17:44:18.566714 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcqt7" event={"ID":"26115b90-b06a-4f5c-8586-6d4cbe3c2869","Type":"ContainerDied","Data":"6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3"} Dec 09 17:44:19 crc kubenswrapper[4954]: I1209 17:44:19.578287 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcqt7" event={"ID":"26115b90-b06a-4f5c-8586-6d4cbe3c2869","Type":"ContainerStarted","Data":"9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49"} Dec 09 17:44:19 crc kubenswrapper[4954]: I1209 17:44:19.604803 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xcqt7" podStartSLOduration=2.144752942 podStartE2EDuration="6.604783207s" podCreationTimestamp="2025-12-09 17:44:13 +0000 UTC" firstStartedPulling="2025-12-09 17:44:14.515456933 +0000 UTC m=+2850.903630753" lastFinishedPulling="2025-12-09 17:44:18.975487188 +0000 UTC m=+2855.363661018" observedRunningTime="2025-12-09 17:44:19.59664914 +0000 UTC m=+2855.984822970" watchObservedRunningTime="2025-12-09 17:44:19.604783207 +0000 UTC m=+2855.992957027" Dec 09 17:44:23 crc kubenswrapper[4954]: E1209 17:44:23.123176 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:44:23 crc kubenswrapper[4954]: I1209 17:44:23.645019 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:23 crc kubenswrapper[4954]: I1209 17:44:23.645072 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:24 crc kubenswrapper[4954]: I1209 17:44:24.690634 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xcqt7" podUID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerName="registry-server" probeResult="failure" output=< Dec 09 17:44:24 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 17:44:24 crc kubenswrapper[4954]: > Dec 09 17:44:25 crc kubenswrapper[4954]: E1209 17:44:25.121880 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:44:33 crc kubenswrapper[4954]: I1209 17:44:33.701266 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:33 crc kubenswrapper[4954]: I1209 17:44:33.761765 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:33 crc kubenswrapper[4954]: I1209 17:44:33.940757 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xcqt7"] Dec 09 17:44:34 crc kubenswrapper[4954]: E1209 17:44:34.129804 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:44:35 crc kubenswrapper[4954]: I1209 17:44:35.719567 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xcqt7" podUID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerName="registry-server" containerID="cri-o://9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49" gracePeriod=2 Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.300265 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.377313 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-catalog-content\") pod \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.377440 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t8hp\" (UniqueName: \"kubernetes.io/projected/26115b90-b06a-4f5c-8586-6d4cbe3c2869-kube-api-access-9t8hp\") pod \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.377522 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-utilities\") pod \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\" (UID: \"26115b90-b06a-4f5c-8586-6d4cbe3c2869\") " Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.378640 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-utilities" (OuterVolumeSpecName: "utilities") pod "26115b90-b06a-4f5c-8586-6d4cbe3c2869" (UID: "26115b90-b06a-4f5c-8586-6d4cbe3c2869"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.384789 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26115b90-b06a-4f5c-8586-6d4cbe3c2869-kube-api-access-9t8hp" (OuterVolumeSpecName: "kube-api-access-9t8hp") pod "26115b90-b06a-4f5c-8586-6d4cbe3c2869" (UID: "26115b90-b06a-4f5c-8586-6d4cbe3c2869"). InnerVolumeSpecName "kube-api-access-9t8hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.480335 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t8hp\" (UniqueName: \"kubernetes.io/projected/26115b90-b06a-4f5c-8586-6d4cbe3c2869-kube-api-access-9t8hp\") on node \"crc\" DevicePath \"\"" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.480369 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.499574 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26115b90-b06a-4f5c-8586-6d4cbe3c2869" (UID: "26115b90-b06a-4f5c-8586-6d4cbe3c2869"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.582713 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26115b90-b06a-4f5c-8586-6d4cbe3c2869-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.736618 4954 generic.go:334] "Generic (PLEG): container finished" podID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerID="9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49" exitCode=0 Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.736695 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcqt7" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.736699 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcqt7" event={"ID":"26115b90-b06a-4f5c-8586-6d4cbe3c2869","Type":"ContainerDied","Data":"9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49"} Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.737035 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcqt7" event={"ID":"26115b90-b06a-4f5c-8586-6d4cbe3c2869","Type":"ContainerDied","Data":"3fd263d89e23b30e16159d8d9c6b2b7b0478a187f23e87281aa4b9a44111cf80"} Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.737072 4954 scope.go:117] "RemoveContainer" containerID="9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.756324 4954 scope.go:117] "RemoveContainer" containerID="6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.774696 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xcqt7"] Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.785414 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xcqt7"] Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.792827 4954 scope.go:117] "RemoveContainer" containerID="90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.831194 4954 scope.go:117] "RemoveContainer" containerID="9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49" Dec 09 17:44:36 crc kubenswrapper[4954]: E1209 17:44:36.831716 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49\": container with ID starting with 9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49 not found: ID does not exist" containerID="9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.831752 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49"} err="failed to get container status \"9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49\": rpc error: code = NotFound desc = could not find container \"9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49\": container with ID starting with 9bc6ea9e61f7c72b938c143e54828de0b19feb555b6ff3cc2b5693bdab4d8b49 not found: ID does not exist" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.831774 4954 scope.go:117] "RemoveContainer" containerID="6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3" Dec 09 17:44:36 crc kubenswrapper[4954]: E1209 17:44:36.832081 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3\": container with ID starting with 6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3 not found: ID does not exist" containerID="6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.832109 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3"} err="failed to get container status \"6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3\": rpc error: code = NotFound desc = could not find container \"6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3\": container with ID starting with 6c15429e20d11fde3c4eae2b2459bd9d24b1cddbaf4d158bcbcd52c50c7fe0b3 not found: ID does not exist" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.832127 4954 scope.go:117] "RemoveContainer" containerID="90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862" Dec 09 17:44:36 crc kubenswrapper[4954]: E1209 17:44:36.832443 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862\": container with ID starting with 90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862 not found: ID does not exist" containerID="90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862" Dec 09 17:44:36 crc kubenswrapper[4954]: I1209 17:44:36.832466 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862"} err="failed to get container status \"90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862\": rpc error: code = NotFound desc = could not find container \"90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862\": container with ID starting with 90f279004ca9b37c8b91c5301efffeb05c80444925eccfc4329fc85f7ac80862 not found: ID does not exist" Dec 09 17:44:37 crc kubenswrapper[4954]: E1209 17:44:37.123627 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:44:38 crc kubenswrapper[4954]: I1209 17:44:38.136925 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" path="/var/lib/kubelet/pods/26115b90-b06a-4f5c-8586-6d4cbe3c2869/volumes" Dec 09 17:44:43 crc kubenswrapper[4954]: I1209 17:44:43.755636 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:44:43 crc kubenswrapper[4954]: I1209 17:44:43.756765 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:44:45 crc kubenswrapper[4954]: E1209 17:44:45.121856 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:44:51 crc kubenswrapper[4954]: E1209 17:44:51.123034 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:45:00 crc kubenswrapper[4954]: E1209 17:45:00.123777 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.151339 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg"] Dec 09 17:45:00 crc kubenswrapper[4954]: E1209 17:45:00.152107 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerName="registry-server" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.152129 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerName="registry-server" Dec 09 17:45:00 crc kubenswrapper[4954]: E1209 17:45:00.152145 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerName="extract-utilities" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.152151 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerName="extract-utilities" Dec 09 17:45:00 crc kubenswrapper[4954]: E1209 17:45:00.152171 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerName="extract-content" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.152177 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerName="extract-content" Dec 09 17:45:00 crc kubenswrapper[4954]: E1209 17:45:00.152201 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerName="extract-utilities" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.152208 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerName="extract-utilities" Dec 09 17:45:00 crc kubenswrapper[4954]: E1209 17:45:00.152223 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerName="extract-content" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.152229 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerName="extract-content" Dec 09 17:45:00 crc kubenswrapper[4954]: E1209 17:45:00.152239 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerName="registry-server" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.152245 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerName="registry-server" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.152445 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="84354081-af02-4bbf-8ad1-9d2cf8ae2621" containerName="registry-server" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.152477 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="26115b90-b06a-4f5c-8586-6d4cbe3c2869" containerName="registry-server" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.153570 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.155916 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.156179 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.164216 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg"] Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.194410 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6882048a-c4d8-4669-a3c0-262dad0a4fed-secret-volume\") pod \"collect-profiles-29421705-hj5wg\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.194759 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq5jm\" (UniqueName: \"kubernetes.io/projected/6882048a-c4d8-4669-a3c0-262dad0a4fed-kube-api-access-tq5jm\") pod \"collect-profiles-29421705-hj5wg\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.194827 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6882048a-c4d8-4669-a3c0-262dad0a4fed-config-volume\") pod \"collect-profiles-29421705-hj5wg\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.297279 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6882048a-c4d8-4669-a3c0-262dad0a4fed-secret-volume\") pod \"collect-profiles-29421705-hj5wg\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.297328 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq5jm\" (UniqueName: \"kubernetes.io/projected/6882048a-c4d8-4669-a3c0-262dad0a4fed-kube-api-access-tq5jm\") pod \"collect-profiles-29421705-hj5wg\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.297363 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6882048a-c4d8-4669-a3c0-262dad0a4fed-config-volume\") pod \"collect-profiles-29421705-hj5wg\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.298509 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6882048a-c4d8-4669-a3c0-262dad0a4fed-config-volume\") pod \"collect-profiles-29421705-hj5wg\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.304436 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6882048a-c4d8-4669-a3c0-262dad0a4fed-secret-volume\") pod \"collect-profiles-29421705-hj5wg\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.320623 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq5jm\" (UniqueName: \"kubernetes.io/projected/6882048a-c4d8-4669-a3c0-262dad0a4fed-kube-api-access-tq5jm\") pod \"collect-profiles-29421705-hj5wg\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.487695 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:00 crc kubenswrapper[4954]: I1209 17:45:00.952664 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg"] Dec 09 17:45:00 crc kubenswrapper[4954]: W1209 17:45:00.956267 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6882048a_c4d8_4669_a3c0_262dad0a4fed.slice/crio-a0b894a6b009f8cd67c4fc9a936b3aa0f1a93302c844e8753c134aa5108106d8 WatchSource:0}: Error finding container a0b894a6b009f8cd67c4fc9a936b3aa0f1a93302c844e8753c134aa5108106d8: Status 404 returned error can't find the container with id a0b894a6b009f8cd67c4fc9a936b3aa0f1a93302c844e8753c134aa5108106d8 Dec 09 17:45:01 crc kubenswrapper[4954]: I1209 17:45:01.000664 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" event={"ID":"6882048a-c4d8-4669-a3c0-262dad0a4fed","Type":"ContainerStarted","Data":"a0b894a6b009f8cd67c4fc9a936b3aa0f1a93302c844e8753c134aa5108106d8"} Dec 09 17:45:02 crc kubenswrapper[4954]: I1209 17:45:02.016325 4954 generic.go:334] "Generic (PLEG): container finished" podID="6882048a-c4d8-4669-a3c0-262dad0a4fed" containerID="9727c9686cd96a73887834781ed2a0ea0ddbe396363e339e075bd0f690b60c2c" exitCode=0 Dec 09 17:45:02 crc kubenswrapper[4954]: I1209 17:45:02.017142 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" event={"ID":"6882048a-c4d8-4669-a3c0-262dad0a4fed","Type":"ContainerDied","Data":"9727c9686cd96a73887834781ed2a0ea0ddbe396363e339e075bd0f690b60c2c"} Dec 09 17:45:03 crc kubenswrapper[4954]: E1209 17:45:03.122709 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:45:03 crc kubenswrapper[4954]: I1209 17:45:03.402836 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:03 crc kubenswrapper[4954]: I1209 17:45:03.466435 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6882048a-c4d8-4669-a3c0-262dad0a4fed-secret-volume\") pod \"6882048a-c4d8-4669-a3c0-262dad0a4fed\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " Dec 09 17:45:03 crc kubenswrapper[4954]: I1209 17:45:03.466663 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6882048a-c4d8-4669-a3c0-262dad0a4fed-config-volume\") pod \"6882048a-c4d8-4669-a3c0-262dad0a4fed\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " Dec 09 17:45:03 crc kubenswrapper[4954]: I1209 17:45:03.466785 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq5jm\" (UniqueName: \"kubernetes.io/projected/6882048a-c4d8-4669-a3c0-262dad0a4fed-kube-api-access-tq5jm\") pod \"6882048a-c4d8-4669-a3c0-262dad0a4fed\" (UID: \"6882048a-c4d8-4669-a3c0-262dad0a4fed\") " Dec 09 17:45:03 crc kubenswrapper[4954]: I1209 17:45:03.467559 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6882048a-c4d8-4669-a3c0-262dad0a4fed-config-volume" (OuterVolumeSpecName: "config-volume") pod "6882048a-c4d8-4669-a3c0-262dad0a4fed" (UID: "6882048a-c4d8-4669-a3c0-262dad0a4fed"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 17:45:03 crc kubenswrapper[4954]: I1209 17:45:03.473790 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6882048a-c4d8-4669-a3c0-262dad0a4fed-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6882048a-c4d8-4669-a3c0-262dad0a4fed" (UID: "6882048a-c4d8-4669-a3c0-262dad0a4fed"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:45:03 crc kubenswrapper[4954]: I1209 17:45:03.473851 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6882048a-c4d8-4669-a3c0-262dad0a4fed-kube-api-access-tq5jm" (OuterVolumeSpecName: "kube-api-access-tq5jm") pod "6882048a-c4d8-4669-a3c0-262dad0a4fed" (UID: "6882048a-c4d8-4669-a3c0-262dad0a4fed"). InnerVolumeSpecName "kube-api-access-tq5jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:45:03 crc kubenswrapper[4954]: I1209 17:45:03.568930 4954 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6882048a-c4d8-4669-a3c0-262dad0a4fed-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 17:45:03 crc kubenswrapper[4954]: I1209 17:45:03.568966 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq5jm\" (UniqueName: \"kubernetes.io/projected/6882048a-c4d8-4669-a3c0-262dad0a4fed-kube-api-access-tq5jm\") on node \"crc\" DevicePath \"\"" Dec 09 17:45:03 crc kubenswrapper[4954]: I1209 17:45:03.568980 4954 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6882048a-c4d8-4669-a3c0-262dad0a4fed-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 17:45:04 crc kubenswrapper[4954]: I1209 17:45:04.041570 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" event={"ID":"6882048a-c4d8-4669-a3c0-262dad0a4fed","Type":"ContainerDied","Data":"a0b894a6b009f8cd67c4fc9a936b3aa0f1a93302c844e8753c134aa5108106d8"} Dec 09 17:45:04 crc kubenswrapper[4954]: I1209 17:45:04.041659 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0b894a6b009f8cd67c4fc9a936b3aa0f1a93302c844e8753c134aa5108106d8" Dec 09 17:45:04 crc kubenswrapper[4954]: I1209 17:45:04.041738 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg" Dec 09 17:45:04 crc kubenswrapper[4954]: I1209 17:45:04.477115 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd"] Dec 09 17:45:04 crc kubenswrapper[4954]: I1209 17:45:04.486334 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421660-fnwcd"] Dec 09 17:45:06 crc kubenswrapper[4954]: I1209 17:45:06.152374 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd" path="/var/lib/kubelet/pods/dfb7d6bd-431a-4dc7-a2e3-d3d4850deabd/volumes" Dec 09 17:45:13 crc kubenswrapper[4954]: E1209 17:45:13.123971 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:45:13 crc kubenswrapper[4954]: I1209 17:45:13.756153 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:45:13 crc kubenswrapper[4954]: I1209 17:45:13.756546 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:45:13 crc kubenswrapper[4954]: I1209 17:45:13.756621 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:45:13 crc kubenswrapper[4954]: I1209 17:45:13.757586 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b51db8880937853b1d1e5ce56a6c02a6aa8e1062a55e02266d9eb62deba3e8f"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:45:13 crc kubenswrapper[4954]: I1209 17:45:13.757711 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://8b51db8880937853b1d1e5ce56a6c02a6aa8e1062a55e02266d9eb62deba3e8f" gracePeriod=600 Dec 09 17:45:14 crc kubenswrapper[4954]: E1209 17:45:14.130882 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:45:14 crc kubenswrapper[4954]: I1209 17:45:14.142485 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="8b51db8880937853b1d1e5ce56a6c02a6aa8e1062a55e02266d9eb62deba3e8f" exitCode=0 Dec 09 17:45:14 crc kubenswrapper[4954]: I1209 17:45:14.142544 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"8b51db8880937853b1d1e5ce56a6c02a6aa8e1062a55e02266d9eb62deba3e8f"} Dec 09 17:45:14 crc kubenswrapper[4954]: I1209 17:45:14.142585 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f"} Dec 09 17:45:14 crc kubenswrapper[4954]: I1209 17:45:14.142683 4954 scope.go:117] "RemoveContainer" containerID="23f564dc7d23a6f9d8767d919ee53fc44fc7d25bcd3dd738a0961024f8c7117b" Dec 09 17:45:24 crc kubenswrapper[4954]: E1209 17:45:24.129835 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:45:26 crc kubenswrapper[4954]: E1209 17:45:26.122498 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:45:35 crc kubenswrapper[4954]: E1209 17:45:35.122779 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:45:40 crc kubenswrapper[4954]: E1209 17:45:40.123270 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:45:48 crc kubenswrapper[4954]: E1209 17:45:48.122848 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:45:52 crc kubenswrapper[4954]: E1209 17:45:52.124402 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:46:00 crc kubenswrapper[4954]: E1209 17:46:00.124013 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:46:05 crc kubenswrapper[4954]: I1209 17:46:05.172135 4954 scope.go:117] "RemoveContainer" containerID="00fff6ceec063b5787c807e6e60ca4a38ec961ab1ea3c8d36d3cb655f87502e1" Dec 09 17:46:08 crc kubenswrapper[4954]: E1209 17:46:08.128851 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:46:13 crc kubenswrapper[4954]: E1209 17:46:13.122390 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:46:17 crc kubenswrapper[4954]: I1209 17:46:17.748390 4954 generic.go:334] "Generic (PLEG): container finished" podID="d64c4820-0dfe-46fa-9275-e9d45bbb268e" containerID="06bebe0dc981564d98ab2770c730bfb639e65078b279977d414f77e4c7422aff" exitCode=0 Dec 09 17:46:17 crc kubenswrapper[4954]: I1209 17:46:17.748481 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" event={"ID":"d64c4820-0dfe-46fa-9275-e9d45bbb268e","Type":"ContainerDied","Data":"06bebe0dc981564d98ab2770c730bfb639e65078b279977d414f77e4c7422aff"} Dec 09 17:46:19 crc kubenswrapper[4954]: E1209 17:46:19.131054 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.231822 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.394232 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-secret-0\") pod \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.394414 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dj8v\" (UniqueName: \"kubernetes.io/projected/d64c4820-0dfe-46fa-9275-e9d45bbb268e-kube-api-access-8dj8v\") pod \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.394539 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-inventory\") pod \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.394575 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-combined-ca-bundle\") pod \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.394633 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-ssh-key\") pod \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\" (UID: \"d64c4820-0dfe-46fa-9275-e9d45bbb268e\") " Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.399469 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d64c4820-0dfe-46fa-9275-e9d45bbb268e" (UID: "d64c4820-0dfe-46fa-9275-e9d45bbb268e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.400450 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64c4820-0dfe-46fa-9275-e9d45bbb268e-kube-api-access-8dj8v" (OuterVolumeSpecName: "kube-api-access-8dj8v") pod "d64c4820-0dfe-46fa-9275-e9d45bbb268e" (UID: "d64c4820-0dfe-46fa-9275-e9d45bbb268e"). InnerVolumeSpecName "kube-api-access-8dj8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.423531 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "d64c4820-0dfe-46fa-9275-e9d45bbb268e" (UID: "d64c4820-0dfe-46fa-9275-e9d45bbb268e"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.424141 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d64c4820-0dfe-46fa-9275-e9d45bbb268e" (UID: "d64c4820-0dfe-46fa-9275-e9d45bbb268e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.429432 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-inventory" (OuterVolumeSpecName: "inventory") pod "d64c4820-0dfe-46fa-9275-e9d45bbb268e" (UID: "d64c4820-0dfe-46fa-9275-e9d45bbb268e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.497439 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dj8v\" (UniqueName: \"kubernetes.io/projected/d64c4820-0dfe-46fa-9275-e9d45bbb268e-kube-api-access-8dj8v\") on node \"crc\" DevicePath \"\"" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.497473 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.497483 4954 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.497491 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.497503 4954 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d64c4820-0dfe-46fa-9275-e9d45bbb268e-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.772060 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" event={"ID":"d64c4820-0dfe-46fa-9275-e9d45bbb268e","Type":"ContainerDied","Data":"1459d24c659da2e0650584d10ecf0dc12c4645620e41e9deae4cc3d10efb3062"} Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.772108 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1459d24c659da2e0650584d10ecf0dc12c4645620e41e9deae4cc3d10efb3062" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.772150 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.857403 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp"] Dec 09 17:46:19 crc kubenswrapper[4954]: E1209 17:46:19.858034 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6882048a-c4d8-4669-a3c0-262dad0a4fed" containerName="collect-profiles" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.858065 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="6882048a-c4d8-4669-a3c0-262dad0a4fed" containerName="collect-profiles" Dec 09 17:46:19 crc kubenswrapper[4954]: E1209 17:46:19.858114 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64c4820-0dfe-46fa-9275-e9d45bbb268e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.858125 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64c4820-0dfe-46fa-9275-e9d45bbb268e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.858381 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="6882048a-c4d8-4669-a3c0-262dad0a4fed" containerName="collect-profiles" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.858420 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64c4820-0dfe-46fa-9275-e9d45bbb268e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.859448 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.863681 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.863790 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.863687 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.864036 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.869365 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:46:19 crc kubenswrapper[4954]: I1209 17:46:19.877798 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp"] Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.007578 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.007691 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.007757 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5vfq\" (UniqueName: \"kubernetes.io/projected/ac8da3e6-984f-4813-bd51-595981f71106-kube-api-access-c5vfq\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.007802 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.007828 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.007955 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.007984 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.110097 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.110157 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.110212 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.110288 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.110338 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5vfq\" (UniqueName: \"kubernetes.io/projected/ac8da3e6-984f-4813-bd51-595981f71106-kube-api-access-c5vfq\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.110398 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.110426 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.114895 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.115002 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.115313 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.116618 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.121095 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.122145 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.134130 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5vfq\" (UniqueName: \"kubernetes.io/projected/ac8da3e6-984f-4813-bd51-595981f71106-kube-api-access-c5vfq\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.185226 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.699575 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp"] Dec 09 17:46:20 crc kubenswrapper[4954]: I1209 17:46:20.783758 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" event={"ID":"ac8da3e6-984f-4813-bd51-595981f71106","Type":"ContainerStarted","Data":"2ce7edaadb6b45e27c7587bcb86cc0797b11c882ec7d9577012c7df7c2358af0"} Dec 09 17:46:21 crc kubenswrapper[4954]: I1209 17:46:21.797068 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" event={"ID":"ac8da3e6-984f-4813-bd51-595981f71106","Type":"ContainerStarted","Data":"34a705b0435513c3168ba4a71d8d2fbc7831c78829a9ec5f24fac2ab084c8340"} Dec 09 17:46:21 crc kubenswrapper[4954]: I1209 17:46:21.822539 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" podStartSLOduration=2.303117542 podStartE2EDuration="2.822507923s" podCreationTimestamp="2025-12-09 17:46:19 +0000 UTC" firstStartedPulling="2025-12-09 17:46:20.701540619 +0000 UTC m=+2977.089714469" lastFinishedPulling="2025-12-09 17:46:21.22093102 +0000 UTC m=+2977.609104850" observedRunningTime="2025-12-09 17:46:21.818882599 +0000 UTC m=+2978.207056419" watchObservedRunningTime="2025-12-09 17:46:21.822507923 +0000 UTC m=+2978.210681763" Dec 09 17:46:26 crc kubenswrapper[4954]: E1209 17:46:26.121734 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:46:32 crc kubenswrapper[4954]: E1209 17:46:32.122634 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:46:37 crc kubenswrapper[4954]: E1209 17:46:37.122743 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:46:47 crc kubenswrapper[4954]: E1209 17:46:47.123312 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:46:50 crc kubenswrapper[4954]: E1209 17:46:50.124176 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:47:02 crc kubenswrapper[4954]: E1209 17:47:02.121766 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:47:05 crc kubenswrapper[4954]: E1209 17:47:05.122530 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:47:15 crc kubenswrapper[4954]: I1209 17:47:15.342470 4954 generic.go:334] "Generic (PLEG): container finished" podID="ac8da3e6-984f-4813-bd51-595981f71106" containerID="34a705b0435513c3168ba4a71d8d2fbc7831c78829a9ec5f24fac2ab084c8340" exitCode=2 Dec 09 17:47:15 crc kubenswrapper[4954]: I1209 17:47:15.342553 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" event={"ID":"ac8da3e6-984f-4813-bd51-595981f71106","Type":"ContainerDied","Data":"34a705b0435513c3168ba4a71d8d2fbc7831c78829a9ec5f24fac2ab084c8340"} Dec 09 17:47:16 crc kubenswrapper[4954]: E1209 17:47:16.123533 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.790762 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.962968 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-2\") pod \"ac8da3e6-984f-4813-bd51-595981f71106\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.963050 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5vfq\" (UniqueName: \"kubernetes.io/projected/ac8da3e6-984f-4813-bd51-595981f71106-kube-api-access-c5vfq\") pod \"ac8da3e6-984f-4813-bd51-595981f71106\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.963140 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ssh-key\") pod \"ac8da3e6-984f-4813-bd51-595981f71106\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.963212 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-1\") pod \"ac8da3e6-984f-4813-bd51-595981f71106\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.963266 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-0\") pod \"ac8da3e6-984f-4813-bd51-595981f71106\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.963301 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-telemetry-combined-ca-bundle\") pod \"ac8da3e6-984f-4813-bd51-595981f71106\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.963337 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-inventory\") pod \"ac8da3e6-984f-4813-bd51-595981f71106\" (UID: \"ac8da3e6-984f-4813-bd51-595981f71106\") " Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.969611 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac8da3e6-984f-4813-bd51-595981f71106-kube-api-access-c5vfq" (OuterVolumeSpecName: "kube-api-access-c5vfq") pod "ac8da3e6-984f-4813-bd51-595981f71106" (UID: "ac8da3e6-984f-4813-bd51-595981f71106"). InnerVolumeSpecName "kube-api-access-c5vfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.969674 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "ac8da3e6-984f-4813-bd51-595981f71106" (UID: "ac8da3e6-984f-4813-bd51-595981f71106"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.998714 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-inventory" (OuterVolumeSpecName: "inventory") pod "ac8da3e6-984f-4813-bd51-595981f71106" (UID: "ac8da3e6-984f-4813-bd51-595981f71106"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:47:16 crc kubenswrapper[4954]: I1209 17:47:16.999934 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "ac8da3e6-984f-4813-bd51-595981f71106" (UID: "ac8da3e6-984f-4813-bd51-595981f71106"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.001751 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ac8da3e6-984f-4813-bd51-595981f71106" (UID: "ac8da3e6-984f-4813-bd51-595981f71106"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.002037 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "ac8da3e6-984f-4813-bd51-595981f71106" (UID: "ac8da3e6-984f-4813-bd51-595981f71106"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.002163 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "ac8da3e6-984f-4813-bd51-595981f71106" (UID: "ac8da3e6-984f-4813-bd51-595981f71106"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.066644 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.066687 4954 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.066705 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.066718 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.066731 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5vfq\" (UniqueName: \"kubernetes.io/projected/ac8da3e6-984f-4813-bd51-595981f71106-kube-api-access-c5vfq\") on node \"crc\" DevicePath \"\"" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.066742 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.066753 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ac8da3e6-984f-4813-bd51-595981f71106-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 09 17:47:17 crc kubenswrapper[4954]: E1209 17:47:17.126384 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.361868 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" event={"ID":"ac8da3e6-984f-4813-bd51-595981f71106","Type":"ContainerDied","Data":"2ce7edaadb6b45e27c7587bcb86cc0797b11c882ec7d9577012c7df7c2358af0"} Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.361909 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ce7edaadb6b45e27c7587bcb86cc0797b11c882ec7d9577012c7df7c2358af0" Dec 09 17:47:17 crc kubenswrapper[4954]: I1209 17:47:17.361962 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.029161 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd"] Dec 09 17:47:25 crc kubenswrapper[4954]: E1209 17:47:25.030430 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8da3e6-984f-4813-bd51-595981f71106" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.030446 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8da3e6-984f-4813-bd51-595981f71106" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.030695 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac8da3e6-984f-4813-bd51-595981f71106" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.031418 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.033241 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.033573 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.033748 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.033936 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.034541 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.058439 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd"] Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.133661 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.133714 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.133805 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.133856 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.133900 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.134024 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.134091 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpldc\" (UniqueName: \"kubernetes.io/projected/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-kube-api-access-gpldc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.235636 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.235705 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.235740 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpldc\" (UniqueName: \"kubernetes.io/projected/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-kube-api-access-gpldc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.235883 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.235912 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.235999 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.236062 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.242312 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.245026 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.245152 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.258131 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.262328 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpldc\" (UniqueName: \"kubernetes.io/projected/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-kube-api-access-gpldc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.263445 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.264078 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.358700 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.959637 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd"] Dec 09 17:47:25 crc kubenswrapper[4954]: W1209 17:47:25.967820 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eb219ed_3f84_41ec_be8c_1e5d5f8c7dfc.slice/crio-476f9b2b44a3ae34dc519225d97ba96bf1fb688212802ca62ee164fb6a6a0e8f WatchSource:0}: Error finding container 476f9b2b44a3ae34dc519225d97ba96bf1fb688212802ca62ee164fb6a6a0e8f: Status 404 returned error can't find the container with id 476f9b2b44a3ae34dc519225d97ba96bf1fb688212802ca62ee164fb6a6a0e8f Dec 09 17:47:25 crc kubenswrapper[4954]: I1209 17:47:25.975225 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 17:47:26 crc kubenswrapper[4954]: I1209 17:47:26.441313 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" event={"ID":"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc","Type":"ContainerStarted","Data":"476f9b2b44a3ae34dc519225d97ba96bf1fb688212802ca62ee164fb6a6a0e8f"} Dec 09 17:47:27 crc kubenswrapper[4954]: I1209 17:47:27.453303 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" event={"ID":"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc","Type":"ContainerStarted","Data":"88ea15a9a5b235b94098b35b50d44bb71329fa431180bc4a6f9122a93f3e98a5"} Dec 09 17:47:27 crc kubenswrapper[4954]: I1209 17:47:27.471543 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" podStartSLOduration=1.776352507 podStartE2EDuration="2.471521312s" podCreationTimestamp="2025-12-09 17:47:25 +0000 UTC" firstStartedPulling="2025-12-09 17:47:25.974956103 +0000 UTC m=+3042.363129923" lastFinishedPulling="2025-12-09 17:47:26.670124908 +0000 UTC m=+3043.058298728" observedRunningTime="2025-12-09 17:47:27.467481285 +0000 UTC m=+3043.855655105" watchObservedRunningTime="2025-12-09 17:47:27.471521312 +0000 UTC m=+3043.859695132" Dec 09 17:47:29 crc kubenswrapper[4954]: E1209 17:47:29.122665 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:47:31 crc kubenswrapper[4954]: E1209 17:47:31.235148 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:47:31 crc kubenswrapper[4954]: E1209 17:47:31.235484 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:47:31 crc kubenswrapper[4954]: E1209 17:47:31.235765 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:47:31 crc kubenswrapper[4954]: E1209 17:47:31.236939 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:47:43 crc kubenswrapper[4954]: E1209 17:47:43.244581 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:47:43 crc kubenswrapper[4954]: E1209 17:47:43.245289 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:47:43 crc kubenswrapper[4954]: E1209 17:47:43.245417 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:47:43 crc kubenswrapper[4954]: E1209 17:47:43.246680 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:47:43 crc kubenswrapper[4954]: I1209 17:47:43.757080 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:47:43 crc kubenswrapper[4954]: I1209 17:47:43.757184 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:47:46 crc kubenswrapper[4954]: E1209 17:47:46.121490 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:47:57 crc kubenswrapper[4954]: E1209 17:47:57.122725 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:47:58 crc kubenswrapper[4954]: E1209 17:47:58.123398 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:48:09 crc kubenswrapper[4954]: E1209 17:48:09.123900 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:48:13 crc kubenswrapper[4954]: E1209 17:48:13.123547 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:48:13 crc kubenswrapper[4954]: I1209 17:48:13.756355 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:48:13 crc kubenswrapper[4954]: I1209 17:48:13.756457 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:48:17 crc kubenswrapper[4954]: I1209 17:48:17.976186 4954 generic.go:334] "Generic (PLEG): container finished" podID="1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" containerID="88ea15a9a5b235b94098b35b50d44bb71329fa431180bc4a6f9122a93f3e98a5" exitCode=2 Dec 09 17:48:17 crc kubenswrapper[4954]: I1209 17:48:17.976265 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" event={"ID":"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc","Type":"ContainerDied","Data":"88ea15a9a5b235b94098b35b50d44bb71329fa431180bc4a6f9122a93f3e98a5"} Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.908847 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.964152 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-1\") pod \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.964270 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-0\") pod \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.964494 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpldc\" (UniqueName: \"kubernetes.io/projected/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-kube-api-access-gpldc\") pod \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.964551 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-inventory\") pod \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.964698 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-2\") pod \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.964891 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-telemetry-combined-ca-bundle\") pod \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.964940 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ssh-key\") pod \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\" (UID: \"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc\") " Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.971931 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-kube-api-access-gpldc" (OuterVolumeSpecName: "kube-api-access-gpldc") pod "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" (UID: "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc"). InnerVolumeSpecName "kube-api-access-gpldc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.972751 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" (UID: "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.996976 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" event={"ID":"1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc","Type":"ContainerDied","Data":"476f9b2b44a3ae34dc519225d97ba96bf1fb688212802ca62ee164fb6a6a0e8f"} Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.997163 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="476f9b2b44a3ae34dc519225d97ba96bf1fb688212802ca62ee164fb6a6a0e8f" Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.997044 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd" Dec 09 17:48:19 crc kubenswrapper[4954]: I1209 17:48:19.997890 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" (UID: "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.005607 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" (UID: "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.016250 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-inventory" (OuterVolumeSpecName: "inventory") pod "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" (UID: "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.017555 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" (UID: "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.034950 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" (UID: "1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.067130 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.067176 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.067195 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpldc\" (UniqueName: \"kubernetes.io/projected/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-kube-api-access-gpldc\") on node \"crc\" DevicePath \"\"" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.067209 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.067220 4954 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.067234 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:48:20 crc kubenswrapper[4954]: I1209 17:48:20.067244 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 09 17:48:24 crc kubenswrapper[4954]: E1209 17:48:24.132477 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:48:28 crc kubenswrapper[4954]: E1209 17:48:28.122337 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:48:36 crc kubenswrapper[4954]: E1209 17:48:36.122505 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.031200 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78"] Dec 09 17:48:37 crc kubenswrapper[4954]: E1209 17:48:37.032144 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.032173 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.032437 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.033404 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.041120 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.041217 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.042761 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.042821 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.043874 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.062880 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78"] Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.132907 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.133010 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.133062 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmpmw\" (UniqueName: \"kubernetes.io/projected/b21d50b3-8256-4f9e-ad87-681ff4111133-kube-api-access-xmpmw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.133097 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.133217 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.133282 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.133378 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.234319 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.234469 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.234536 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.234566 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmpmw\" (UniqueName: \"kubernetes.io/projected/b21d50b3-8256-4f9e-ad87-681ff4111133-kube-api-access-xmpmw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.234613 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.234682 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.234744 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.240745 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.241535 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.242370 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.242441 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.243636 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.248051 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.251494 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmpmw\" (UniqueName: \"kubernetes.io/projected/b21d50b3-8256-4f9e-ad87-681ff4111133-kube-api-access-xmpmw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-6lr78\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.363261 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:48:37 crc kubenswrapper[4954]: I1209 17:48:37.964449 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78"] Dec 09 17:48:38 crc kubenswrapper[4954]: I1209 17:48:38.176982 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" event={"ID":"b21d50b3-8256-4f9e-ad87-681ff4111133","Type":"ContainerStarted","Data":"ed80cb955885d47f04d2ef3b53adc2a48812393b4345398b024e70f114c07594"} Dec 09 17:48:39 crc kubenswrapper[4954]: E1209 17:48:39.125559 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:48:41 crc kubenswrapper[4954]: I1209 17:48:41.228403 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" event={"ID":"b21d50b3-8256-4f9e-ad87-681ff4111133","Type":"ContainerStarted","Data":"88725ecb3cf3f1f7956cc2170267fc7e4c6f2aa0ac3e5104f8be0dfb768ab464"} Dec 09 17:48:41 crc kubenswrapper[4954]: I1209 17:48:41.260006 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" podStartSLOduration=1.491916555 podStartE2EDuration="4.259973225s" podCreationTimestamp="2025-12-09 17:48:37 +0000 UTC" firstStartedPulling="2025-12-09 17:48:37.96792083 +0000 UTC m=+3114.356094650" lastFinishedPulling="2025-12-09 17:48:40.73597749 +0000 UTC m=+3117.124151320" observedRunningTime="2025-12-09 17:48:41.247701839 +0000 UTC m=+3117.635875679" watchObservedRunningTime="2025-12-09 17:48:41.259973225 +0000 UTC m=+3117.648147055" Dec 09 17:48:43 crc kubenswrapper[4954]: I1209 17:48:43.755501 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:48:43 crc kubenswrapper[4954]: I1209 17:48:43.755999 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:48:43 crc kubenswrapper[4954]: I1209 17:48:43.756045 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:48:43 crc kubenswrapper[4954]: I1209 17:48:43.756871 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:48:43 crc kubenswrapper[4954]: I1209 17:48:43.756930 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" gracePeriod=600 Dec 09 17:48:43 crc kubenswrapper[4954]: E1209 17:48:43.875436 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:48:44 crc kubenswrapper[4954]: I1209 17:48:44.261181 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" exitCode=0 Dec 09 17:48:44 crc kubenswrapper[4954]: I1209 17:48:44.261225 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f"} Dec 09 17:48:44 crc kubenswrapper[4954]: I1209 17:48:44.261939 4954 scope.go:117] "RemoveContainer" containerID="8b51db8880937853b1d1e5ce56a6c02a6aa8e1062a55e02266d9eb62deba3e8f" Dec 09 17:48:44 crc kubenswrapper[4954]: I1209 17:48:44.263112 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:48:44 crc kubenswrapper[4954]: E1209 17:48:44.263617 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:48:49 crc kubenswrapper[4954]: E1209 17:48:49.122723 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:48:50 crc kubenswrapper[4954]: E1209 17:48:50.121164 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:48:57 crc kubenswrapper[4954]: I1209 17:48:57.120450 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:48:57 crc kubenswrapper[4954]: E1209 17:48:57.121397 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:49:00 crc kubenswrapper[4954]: E1209 17:49:00.123179 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:49:05 crc kubenswrapper[4954]: E1209 17:49:05.125898 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:49:12 crc kubenswrapper[4954]: I1209 17:49:12.120145 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:49:12 crc kubenswrapper[4954]: E1209 17:49:12.120903 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:49:14 crc kubenswrapper[4954]: E1209 17:49:14.133548 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:49:17 crc kubenswrapper[4954]: E1209 17:49:17.121823 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:49:23 crc kubenswrapper[4954]: I1209 17:49:23.121445 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:49:23 crc kubenswrapper[4954]: E1209 17:49:23.123177 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:49:27 crc kubenswrapper[4954]: E1209 17:49:27.122861 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:49:27 crc kubenswrapper[4954]: I1209 17:49:27.486064 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-547f8767c7-hv854" podUID="a6635c8e-d9a5-4034-ab5c-7fe96dc10c10" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 09 17:49:29 crc kubenswrapper[4954]: E1209 17:49:29.123300 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:49:31 crc kubenswrapper[4954]: I1209 17:49:31.694617 4954 generic.go:334] "Generic (PLEG): container finished" podID="b21d50b3-8256-4f9e-ad87-681ff4111133" containerID="88725ecb3cf3f1f7956cc2170267fc7e4c6f2aa0ac3e5104f8be0dfb768ab464" exitCode=2 Dec 09 17:49:31 crc kubenswrapper[4954]: I1209 17:49:31.694719 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" event={"ID":"b21d50b3-8256-4f9e-ad87-681ff4111133","Type":"ContainerDied","Data":"88725ecb3cf3f1f7956cc2170267fc7e4c6f2aa0ac3e5104f8be0dfb768ab464"} Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.201192 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.327452 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-2\") pod \"b21d50b3-8256-4f9e-ad87-681ff4111133\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.327544 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ssh-key\") pod \"b21d50b3-8256-4f9e-ad87-681ff4111133\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.327610 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmpmw\" (UniqueName: \"kubernetes.io/projected/b21d50b3-8256-4f9e-ad87-681ff4111133-kube-api-access-xmpmw\") pod \"b21d50b3-8256-4f9e-ad87-681ff4111133\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.327673 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-inventory\") pod \"b21d50b3-8256-4f9e-ad87-681ff4111133\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.327718 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-telemetry-combined-ca-bundle\") pod \"b21d50b3-8256-4f9e-ad87-681ff4111133\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.328074 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-0\") pod \"b21d50b3-8256-4f9e-ad87-681ff4111133\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.328144 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-1\") pod \"b21d50b3-8256-4f9e-ad87-681ff4111133\" (UID: \"b21d50b3-8256-4f9e-ad87-681ff4111133\") " Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.340233 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b21d50b3-8256-4f9e-ad87-681ff4111133" (UID: "b21d50b3-8256-4f9e-ad87-681ff4111133"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.340434 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b21d50b3-8256-4f9e-ad87-681ff4111133-kube-api-access-xmpmw" (OuterVolumeSpecName: "kube-api-access-xmpmw") pod "b21d50b3-8256-4f9e-ad87-681ff4111133" (UID: "b21d50b3-8256-4f9e-ad87-681ff4111133"). InnerVolumeSpecName "kube-api-access-xmpmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.357549 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-inventory" (OuterVolumeSpecName: "inventory") pod "b21d50b3-8256-4f9e-ad87-681ff4111133" (UID: "b21d50b3-8256-4f9e-ad87-681ff4111133"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.358775 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b21d50b3-8256-4f9e-ad87-681ff4111133" (UID: "b21d50b3-8256-4f9e-ad87-681ff4111133"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.361407 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "b21d50b3-8256-4f9e-ad87-681ff4111133" (UID: "b21d50b3-8256-4f9e-ad87-681ff4111133"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.362925 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "b21d50b3-8256-4f9e-ad87-681ff4111133" (UID: "b21d50b3-8256-4f9e-ad87-681ff4111133"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.363741 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "b21d50b3-8256-4f9e-ad87-681ff4111133" (UID: "b21d50b3-8256-4f9e-ad87-681ff4111133"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.431034 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.431068 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.431080 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.431089 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmpmw\" (UniqueName: \"kubernetes.io/projected/b21d50b3-8256-4f9e-ad87-681ff4111133-kube-api-access-xmpmw\") on node \"crc\" DevicePath \"\"" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.431099 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.431110 4954 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.431119 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b21d50b3-8256-4f9e-ad87-681ff4111133-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.715630 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" event={"ID":"b21d50b3-8256-4f9e-ad87-681ff4111133","Type":"ContainerDied","Data":"ed80cb955885d47f04d2ef3b53adc2a48812393b4345398b024e70f114c07594"} Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.715682 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed80cb955885d47f04d2ef3b53adc2a48812393b4345398b024e70f114c07594" Dec 09 17:49:33 crc kubenswrapper[4954]: I1209 17:49:33.715711 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-6lr78" Dec 09 17:49:36 crc kubenswrapper[4954]: I1209 17:49:36.121172 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:49:36 crc kubenswrapper[4954]: E1209 17:49:36.121799 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:49:40 crc kubenswrapper[4954]: E1209 17:49:40.122417 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:49:43 crc kubenswrapper[4954]: E1209 17:49:43.122549 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:49:50 crc kubenswrapper[4954]: I1209 17:49:50.120900 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:49:50 crc kubenswrapper[4954]: E1209 17:49:50.121693 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:49:54 crc kubenswrapper[4954]: E1209 17:49:54.143351 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:49:55 crc kubenswrapper[4954]: E1209 17:49:55.122671 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:50:05 crc kubenswrapper[4954]: I1209 17:50:05.120420 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:50:05 crc kubenswrapper[4954]: E1209 17:50:05.121148 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:50:05 crc kubenswrapper[4954]: E1209 17:50:05.122243 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.039583 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd"] Dec 09 17:50:10 crc kubenswrapper[4954]: E1209 17:50:10.044426 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b21d50b3-8256-4f9e-ad87-681ff4111133" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.044457 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b21d50b3-8256-4f9e-ad87-681ff4111133" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.044715 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="b21d50b3-8256-4f9e-ad87-681ff4111133" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.045724 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.054998 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.055285 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.055400 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.055545 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.056025 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.058370 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd"] Dec 09 17:50:10 crc kubenswrapper[4954]: E1209 17:50:10.123199 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.195317 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.196001 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.196951 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.197142 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.197186 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ss52\" (UniqueName: \"kubernetes.io/projected/31ae7df8-13b6-4691-8d2b-3b5880921353-kube-api-access-5ss52\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.197281 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.197681 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.300706 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.301693 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.301799 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.301918 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.301978 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.301999 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ss52\" (UniqueName: \"kubernetes.io/projected/31ae7df8-13b6-4691-8d2b-3b5880921353-kube-api-access-5ss52\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.302023 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.308075 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.308341 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.308989 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.309280 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.309567 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.310206 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.321460 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ss52\" (UniqueName: \"kubernetes.io/projected/31ae7df8-13b6-4691-8d2b-3b5880921353-kube-api-access-5ss52\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-87vhd\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.374670 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:50:10 crc kubenswrapper[4954]: I1209 17:50:10.901392 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd"] Dec 09 17:50:11 crc kubenswrapper[4954]: I1209 17:50:11.255109 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" event={"ID":"31ae7df8-13b6-4691-8d2b-3b5880921353","Type":"ContainerStarted","Data":"b87cb60e6fb9cb1eef68f7bb526bf036f05ce0a0e4435753327fe8f862839c91"} Dec 09 17:50:12 crc kubenswrapper[4954]: I1209 17:50:12.277356 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" event={"ID":"31ae7df8-13b6-4691-8d2b-3b5880921353","Type":"ContainerStarted","Data":"d58b2696b8e7cec4edcc85480c9baf9e31c5404cc3a4be652bcc7b83f4130828"} Dec 09 17:50:12 crc kubenswrapper[4954]: I1209 17:50:12.315400 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" podStartSLOduration=1.794073985 podStartE2EDuration="2.315377246s" podCreationTimestamp="2025-12-09 17:50:10 +0000 UTC" firstStartedPulling="2025-12-09 17:50:10.900305866 +0000 UTC m=+3207.288479686" lastFinishedPulling="2025-12-09 17:50:11.421609127 +0000 UTC m=+3207.809782947" observedRunningTime="2025-12-09 17:50:12.303407068 +0000 UTC m=+3208.691580908" watchObservedRunningTime="2025-12-09 17:50:12.315377246 +0000 UTC m=+3208.703551066" Dec 09 17:50:16 crc kubenswrapper[4954]: I1209 17:50:16.121731 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:50:16 crc kubenswrapper[4954]: E1209 17:50:16.122645 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:50:19 crc kubenswrapper[4954]: E1209 17:50:19.122578 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:50:21 crc kubenswrapper[4954]: E1209 17:50:21.122504 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:50:30 crc kubenswrapper[4954]: I1209 17:50:30.121052 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:50:30 crc kubenswrapper[4954]: E1209 17:50:30.121954 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:50:32 crc kubenswrapper[4954]: E1209 17:50:32.124958 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:50:35 crc kubenswrapper[4954]: E1209 17:50:35.122298 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:50:42 crc kubenswrapper[4954]: I1209 17:50:42.120912 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:50:42 crc kubenswrapper[4954]: E1209 17:50:42.122063 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:50:44 crc kubenswrapper[4954]: E1209 17:50:44.133569 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:50:47 crc kubenswrapper[4954]: E1209 17:50:47.125892 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:50:56 crc kubenswrapper[4954]: I1209 17:50:56.121357 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:50:56 crc kubenswrapper[4954]: E1209 17:50:56.122286 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:50:56 crc kubenswrapper[4954]: E1209 17:50:56.122958 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:50:58 crc kubenswrapper[4954]: E1209 17:50:58.121636 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.732589 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pm8wb"] Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.736057 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.746243 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pm8wb"] Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.802091 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-utilities\") pod \"community-operators-pm8wb\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.802196 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdzqq\" (UniqueName: \"kubernetes.io/projected/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-kube-api-access-mdzqq\") pod \"community-operators-pm8wb\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.802234 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-catalog-content\") pod \"community-operators-pm8wb\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.904535 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-utilities\") pod \"community-operators-pm8wb\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.904912 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdzqq\" (UniqueName: \"kubernetes.io/projected/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-kube-api-access-mdzqq\") pod \"community-operators-pm8wb\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.904985 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-catalog-content\") pod \"community-operators-pm8wb\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.904992 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-utilities\") pod \"community-operators-pm8wb\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.905473 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-catalog-content\") pod \"community-operators-pm8wb\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:58 crc kubenswrapper[4954]: I1209 17:50:58.928385 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdzqq\" (UniqueName: \"kubernetes.io/projected/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-kube-api-access-mdzqq\") pod \"community-operators-pm8wb\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:59 crc kubenswrapper[4954]: I1209 17:50:59.104337 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:50:59 crc kubenswrapper[4954]: I1209 17:50:59.626423 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pm8wb"] Dec 09 17:50:59 crc kubenswrapper[4954]: I1209 17:50:59.744466 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm8wb" event={"ID":"83d2f5c1-3e9f-400c-84d5-e5276a6384f8","Type":"ContainerStarted","Data":"63b50477531d50ce727b2fdf5fd459a97f6b89723e484d591adc11c0b36324f3"} Dec 09 17:51:00 crc kubenswrapper[4954]: I1209 17:51:00.756974 4954 generic.go:334] "Generic (PLEG): container finished" podID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerID="689e8eae7865a47cbab6a244c5f8b2bb5e5806a20c53ff7fe81d377c22c8f412" exitCode=0 Dec 09 17:51:00 crc kubenswrapper[4954]: I1209 17:51:00.757080 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm8wb" event={"ID":"83d2f5c1-3e9f-400c-84d5-e5276a6384f8","Type":"ContainerDied","Data":"689e8eae7865a47cbab6a244c5f8b2bb5e5806a20c53ff7fe81d377c22c8f412"} Dec 09 17:51:00 crc kubenswrapper[4954]: I1209 17:51:00.760924 4954 generic.go:334] "Generic (PLEG): container finished" podID="31ae7df8-13b6-4691-8d2b-3b5880921353" containerID="d58b2696b8e7cec4edcc85480c9baf9e31c5404cc3a4be652bcc7b83f4130828" exitCode=2 Dec 09 17:51:00 crc kubenswrapper[4954]: I1209 17:51:00.760959 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" event={"ID":"31ae7df8-13b6-4691-8d2b-3b5880921353","Type":"ContainerDied","Data":"d58b2696b8e7cec4edcc85480c9baf9e31c5404cc3a4be652bcc7b83f4130828"} Dec 09 17:51:01 crc kubenswrapper[4954]: I1209 17:51:01.773121 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm8wb" event={"ID":"83d2f5c1-3e9f-400c-84d5-e5276a6384f8","Type":"ContainerStarted","Data":"f2115e5c7a81ff6b6c6ac882e03bb015308783388af6a6b81834a906d9051618"} Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.220686 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.274729 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-inventory\") pod \"31ae7df8-13b6-4691-8d2b-3b5880921353\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.274795 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-1\") pod \"31ae7df8-13b6-4691-8d2b-3b5880921353\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.274825 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ssh-key\") pod \"31ae7df8-13b6-4691-8d2b-3b5880921353\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.274856 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-telemetry-combined-ca-bundle\") pod \"31ae7df8-13b6-4691-8d2b-3b5880921353\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.274917 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-0\") pod \"31ae7df8-13b6-4691-8d2b-3b5880921353\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.275061 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ss52\" (UniqueName: \"kubernetes.io/projected/31ae7df8-13b6-4691-8d2b-3b5880921353-kube-api-access-5ss52\") pod \"31ae7df8-13b6-4691-8d2b-3b5880921353\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.275100 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-2\") pod \"31ae7df8-13b6-4691-8d2b-3b5880921353\" (UID: \"31ae7df8-13b6-4691-8d2b-3b5880921353\") " Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.282150 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31ae7df8-13b6-4691-8d2b-3b5880921353-kube-api-access-5ss52" (OuterVolumeSpecName: "kube-api-access-5ss52") pod "31ae7df8-13b6-4691-8d2b-3b5880921353" (UID: "31ae7df8-13b6-4691-8d2b-3b5880921353"). InnerVolumeSpecName "kube-api-access-5ss52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.293956 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "31ae7df8-13b6-4691-8d2b-3b5880921353" (UID: "31ae7df8-13b6-4691-8d2b-3b5880921353"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.309197 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "31ae7df8-13b6-4691-8d2b-3b5880921353" (UID: "31ae7df8-13b6-4691-8d2b-3b5880921353"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.318452 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "31ae7df8-13b6-4691-8d2b-3b5880921353" (UID: "31ae7df8-13b6-4691-8d2b-3b5880921353"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.320144 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "31ae7df8-13b6-4691-8d2b-3b5880921353" (UID: "31ae7df8-13b6-4691-8d2b-3b5880921353"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.321090 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-inventory" (OuterVolumeSpecName: "inventory") pod "31ae7df8-13b6-4691-8d2b-3b5880921353" (UID: "31ae7df8-13b6-4691-8d2b-3b5880921353"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.326272 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "31ae7df8-13b6-4691-8d2b-3b5880921353" (UID: "31ae7df8-13b6-4691-8d2b-3b5880921353"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.377973 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.378279 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.378292 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.378302 4954 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.378313 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.378324 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ss52\" (UniqueName: \"kubernetes.io/projected/31ae7df8-13b6-4691-8d2b-3b5880921353-kube-api-access-5ss52\") on node \"crc\" DevicePath \"\"" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.378337 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/31ae7df8-13b6-4691-8d2b-3b5880921353-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.785162 4954 generic.go:334] "Generic (PLEG): container finished" podID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerID="f2115e5c7a81ff6b6c6ac882e03bb015308783388af6a6b81834a906d9051618" exitCode=0 Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.785246 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm8wb" event={"ID":"83d2f5c1-3e9f-400c-84d5-e5276a6384f8","Type":"ContainerDied","Data":"f2115e5c7a81ff6b6c6ac882e03bb015308783388af6a6b81834a906d9051618"} Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.788230 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" event={"ID":"31ae7df8-13b6-4691-8d2b-3b5880921353","Type":"ContainerDied","Data":"b87cb60e6fb9cb1eef68f7bb526bf036f05ce0a0e4435753327fe8f862839c91"} Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.788272 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-87vhd" Dec 09 17:51:02 crc kubenswrapper[4954]: I1209 17:51:02.788287 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b87cb60e6fb9cb1eef68f7bb526bf036f05ce0a0e4435753327fe8f862839c91" Dec 09 17:51:04 crc kubenswrapper[4954]: I1209 17:51:04.807932 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm8wb" event={"ID":"83d2f5c1-3e9f-400c-84d5-e5276a6384f8","Type":"ContainerStarted","Data":"4cc9f45c0cd874169c88aae1965894bb44264fa653f2fe7d0828debbd697e39e"} Dec 09 17:51:04 crc kubenswrapper[4954]: I1209 17:51:04.844189 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pm8wb" podStartSLOduration=3.384480088 podStartE2EDuration="6.844165701s" podCreationTimestamp="2025-12-09 17:50:58 +0000 UTC" firstStartedPulling="2025-12-09 17:51:00.760461548 +0000 UTC m=+3257.148635368" lastFinishedPulling="2025-12-09 17:51:04.220147161 +0000 UTC m=+3260.608320981" observedRunningTime="2025-12-09 17:51:04.834424415 +0000 UTC m=+3261.222598245" watchObservedRunningTime="2025-12-09 17:51:04.844165701 +0000 UTC m=+3261.232339521" Dec 09 17:51:07 crc kubenswrapper[4954]: I1209 17:51:07.121111 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:51:07 crc kubenswrapper[4954]: E1209 17:51:07.122248 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:51:07 crc kubenswrapper[4954]: E1209 17:51:07.135562 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:51:09 crc kubenswrapper[4954]: I1209 17:51:09.105388 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:51:09 crc kubenswrapper[4954]: I1209 17:51:09.105817 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:51:09 crc kubenswrapper[4954]: I1209 17:51:09.155637 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:51:09 crc kubenswrapper[4954]: I1209 17:51:09.909511 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:51:09 crc kubenswrapper[4954]: I1209 17:51:09.959712 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pm8wb"] Dec 09 17:51:10 crc kubenswrapper[4954]: E1209 17:51:10.122057 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:51:11 crc kubenswrapper[4954]: I1209 17:51:11.880435 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pm8wb" podUID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerName="registry-server" containerID="cri-o://4cc9f45c0cd874169c88aae1965894bb44264fa653f2fe7d0828debbd697e39e" gracePeriod=2 Dec 09 17:51:12 crc kubenswrapper[4954]: I1209 17:51:12.901116 4954 generic.go:334] "Generic (PLEG): container finished" podID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerID="4cc9f45c0cd874169c88aae1965894bb44264fa653f2fe7d0828debbd697e39e" exitCode=0 Dec 09 17:51:12 crc kubenswrapper[4954]: I1209 17:51:12.901213 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm8wb" event={"ID":"83d2f5c1-3e9f-400c-84d5-e5276a6384f8","Type":"ContainerDied","Data":"4cc9f45c0cd874169c88aae1965894bb44264fa653f2fe7d0828debbd697e39e"} Dec 09 17:51:12 crc kubenswrapper[4954]: I1209 17:51:12.901585 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm8wb" event={"ID":"83d2f5c1-3e9f-400c-84d5-e5276a6384f8","Type":"ContainerDied","Data":"63b50477531d50ce727b2fdf5fd459a97f6b89723e484d591adc11c0b36324f3"} Dec 09 17:51:12 crc kubenswrapper[4954]: I1209 17:51:12.901617 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63b50477531d50ce727b2fdf5fd459a97f6b89723e484d591adc11c0b36324f3" Dec 09 17:51:12 crc kubenswrapper[4954]: I1209 17:51:12.964492 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.021124 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-catalog-content\") pod \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.021391 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-utilities\") pod \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.021435 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdzqq\" (UniqueName: \"kubernetes.io/projected/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-kube-api-access-mdzqq\") pod \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\" (UID: \"83d2f5c1-3e9f-400c-84d5-e5276a6384f8\") " Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.022175 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-utilities" (OuterVolumeSpecName: "utilities") pod "83d2f5c1-3e9f-400c-84d5-e5276a6384f8" (UID: "83d2f5c1-3e9f-400c-84d5-e5276a6384f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.027825 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-kube-api-access-mdzqq" (OuterVolumeSpecName: "kube-api-access-mdzqq") pod "83d2f5c1-3e9f-400c-84d5-e5276a6384f8" (UID: "83d2f5c1-3e9f-400c-84d5-e5276a6384f8"). InnerVolumeSpecName "kube-api-access-mdzqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.080782 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83d2f5c1-3e9f-400c-84d5-e5276a6384f8" (UID: "83d2f5c1-3e9f-400c-84d5-e5276a6384f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.124008 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.124054 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.124066 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdzqq\" (UniqueName: \"kubernetes.io/projected/83d2f5c1-3e9f-400c-84d5-e5276a6384f8-kube-api-access-mdzqq\") on node \"crc\" DevicePath \"\"" Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.910362 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pm8wb" Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.948184 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pm8wb"] Dec 09 17:51:13 crc kubenswrapper[4954]: I1209 17:51:13.957824 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pm8wb"] Dec 09 17:51:14 crc kubenswrapper[4954]: I1209 17:51:14.133776 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" path="/var/lib/kubelet/pods/83d2f5c1-3e9f-400c-84d5-e5276a6384f8/volumes" Dec 09 17:51:19 crc kubenswrapper[4954]: E1209 17:51:19.122416 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:51:20 crc kubenswrapper[4954]: I1209 17:51:20.121468 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:51:20 crc kubenswrapper[4954]: E1209 17:51:20.122180 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:51:21 crc kubenswrapper[4954]: E1209 17:51:21.121406 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:51:32 crc kubenswrapper[4954]: I1209 17:51:32.120880 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:51:32 crc kubenswrapper[4954]: E1209 17:51:32.121787 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:51:32 crc kubenswrapper[4954]: E1209 17:51:32.123304 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:51:36 crc kubenswrapper[4954]: E1209 17:51:36.122529 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:51:43 crc kubenswrapper[4954]: E1209 17:51:43.122422 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:51:45 crc kubenswrapper[4954]: I1209 17:51:45.156265 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:51:45 crc kubenswrapper[4954]: E1209 17:51:45.156836 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:51:48 crc kubenswrapper[4954]: E1209 17:51:48.123063 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:51:57 crc kubenswrapper[4954]: E1209 17:51:57.122603 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:51:59 crc kubenswrapper[4954]: I1209 17:51:59.120677 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:51:59 crc kubenswrapper[4954]: E1209 17:51:59.121277 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:51:59 crc kubenswrapper[4954]: E1209 17:51:59.122655 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:52:10 crc kubenswrapper[4954]: I1209 17:52:10.120680 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:52:10 crc kubenswrapper[4954]: E1209 17:52:10.121494 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:52:12 crc kubenswrapper[4954]: E1209 17:52:12.123187 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:52:13 crc kubenswrapper[4954]: E1209 17:52:13.123125 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.041174 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl"] Dec 09 17:52:20 crc kubenswrapper[4954]: E1209 17:52:20.042924 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerName="extract-content" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.042945 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerName="extract-content" Dec 09 17:52:20 crc kubenswrapper[4954]: E1209 17:52:20.042970 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31ae7df8-13b6-4691-8d2b-3b5880921353" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.042979 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="31ae7df8-13b6-4691-8d2b-3b5880921353" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:52:20 crc kubenswrapper[4954]: E1209 17:52:20.043009 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerName="registry-server" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.043016 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerName="registry-server" Dec 09 17:52:20 crc kubenswrapper[4954]: E1209 17:52:20.043059 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerName="extract-utilities" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.043071 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerName="extract-utilities" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.043332 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d2f5c1-3e9f-400c-84d5-e5276a6384f8" containerName="registry-server" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.043364 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="31ae7df8-13b6-4691-8d2b-3b5880921353" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.044562 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.048559 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.048712 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.048746 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.048572 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.049041 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.057140 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl"] Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.142498 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbcr8\" (UniqueName: \"kubernetes.io/projected/0575ea51-efad-4a28-9cf1-b2b2653312d1-kube-api-access-jbcr8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.142629 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.142920 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.142973 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.143002 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.143092 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.143128 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.245917 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.246005 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.246204 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.246292 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.246368 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbcr8\" (UniqueName: \"kubernetes.io/projected/0575ea51-efad-4a28-9cf1-b2b2653312d1-kube-api-access-jbcr8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.246434 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.246611 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.253918 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.255392 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.255711 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.255881 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.256242 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.256339 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.269032 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbcr8\" (UniqueName: \"kubernetes.io/projected/0575ea51-efad-4a28-9cf1-b2b2653312d1-kube-api-access-jbcr8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.371046 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:52:20 crc kubenswrapper[4954]: I1209 17:52:20.994056 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl"] Dec 09 17:52:21 crc kubenswrapper[4954]: I1209 17:52:21.588223 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" event={"ID":"0575ea51-efad-4a28-9cf1-b2b2653312d1","Type":"ContainerStarted","Data":"752fefb1eadf8fd27c67518098a268d77d2f48f56171cb417f60c2852ab47be0"} Dec 09 17:52:22 crc kubenswrapper[4954]: I1209 17:52:22.601219 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" event={"ID":"0575ea51-efad-4a28-9cf1-b2b2653312d1","Type":"ContainerStarted","Data":"a54d3639ba9f450cbc3f77a6d92266fed780e716bc36d641deaffcfae7e140a8"} Dec 09 17:52:22 crc kubenswrapper[4954]: I1209 17:52:22.633721 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" podStartSLOduration=1.812357343 podStartE2EDuration="2.633702938s" podCreationTimestamp="2025-12-09 17:52:20 +0000 UTC" firstStartedPulling="2025-12-09 17:52:20.99260894 +0000 UTC m=+3337.380782760" lastFinishedPulling="2025-12-09 17:52:21.813954535 +0000 UTC m=+3338.202128355" observedRunningTime="2025-12-09 17:52:22.623333011 +0000 UTC m=+3339.011506841" watchObservedRunningTime="2025-12-09 17:52:22.633702938 +0000 UTC m=+3339.021876758" Dec 09 17:52:23 crc kubenswrapper[4954]: I1209 17:52:23.119900 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:52:23 crc kubenswrapper[4954]: E1209 17:52:23.120267 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:52:24 crc kubenswrapper[4954]: E1209 17:52:24.129709 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:52:25 crc kubenswrapper[4954]: E1209 17:52:25.121500 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:52:35 crc kubenswrapper[4954]: I1209 17:52:35.122448 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 17:52:35 crc kubenswrapper[4954]: E1209 17:52:35.247061 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:52:35 crc kubenswrapper[4954]: E1209 17:52:35.247142 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:52:35 crc kubenswrapper[4954]: E1209 17:52:35.247316 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:52:35 crc kubenswrapper[4954]: E1209 17:52:35.248836 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:52:37 crc kubenswrapper[4954]: I1209 17:52:37.120621 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:52:37 crc kubenswrapper[4954]: E1209 17:52:37.121529 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:52:39 crc kubenswrapper[4954]: E1209 17:52:39.122935 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:52:46 crc kubenswrapper[4954]: E1209 17:52:46.123660 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:52:49 crc kubenswrapper[4954]: I1209 17:52:49.120867 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:52:49 crc kubenswrapper[4954]: E1209 17:52:49.122206 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:52:54 crc kubenswrapper[4954]: E1209 17:52:54.249276 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:52:54 crc kubenswrapper[4954]: E1209 17:52:54.250392 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:52:54 crc kubenswrapper[4954]: E1209 17:52:54.250549 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:52:54 crc kubenswrapper[4954]: E1209 17:52:54.251997 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:52:58 crc kubenswrapper[4954]: E1209 17:52:58.122907 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:53:03 crc kubenswrapper[4954]: I1209 17:53:03.120154 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:53:03 crc kubenswrapper[4954]: E1209 17:53:03.122260 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:53:05 crc kubenswrapper[4954]: E1209 17:53:05.122747 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:53:10 crc kubenswrapper[4954]: E1209 17:53:10.122217 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.337680 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9csc9"] Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.340993 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.356406 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9csc9"] Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.427914 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn6hk\" (UniqueName: \"kubernetes.io/projected/534791d6-9c49-4e40-8269-b56d5bac5221-kube-api-access-qn6hk\") pod \"redhat-marketplace-9csc9\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.428216 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-utilities\") pod \"redhat-marketplace-9csc9\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.428670 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-catalog-content\") pod \"redhat-marketplace-9csc9\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.531022 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn6hk\" (UniqueName: \"kubernetes.io/projected/534791d6-9c49-4e40-8269-b56d5bac5221-kube-api-access-qn6hk\") pod \"redhat-marketplace-9csc9\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.531203 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-utilities\") pod \"redhat-marketplace-9csc9\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.531489 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-catalog-content\") pod \"redhat-marketplace-9csc9\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.532315 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-catalog-content\") pod \"redhat-marketplace-9csc9\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.535814 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-utilities\") pod \"redhat-marketplace-9csc9\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.584337 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn6hk\" (UniqueName: \"kubernetes.io/projected/534791d6-9c49-4e40-8269-b56d5bac5221-kube-api-access-qn6hk\") pod \"redhat-marketplace-9csc9\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:11 crc kubenswrapper[4954]: I1209 17:53:11.662791 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:12 crc kubenswrapper[4954]: I1209 17:53:12.111144 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9csc9"] Dec 09 17:53:12 crc kubenswrapper[4954]: W1209 17:53:12.113999 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod534791d6_9c49_4e40_8269_b56d5bac5221.slice/crio-87f3dd5bcb966952c0a0acdc4b70f9e97dcb71b8b444c120129749ab2dfa0e80 WatchSource:0}: Error finding container 87f3dd5bcb966952c0a0acdc4b70f9e97dcb71b8b444c120129749ab2dfa0e80: Status 404 returned error can't find the container with id 87f3dd5bcb966952c0a0acdc4b70f9e97dcb71b8b444c120129749ab2dfa0e80 Dec 09 17:53:13 crc kubenswrapper[4954]: I1209 17:53:13.124982 4954 generic.go:334] "Generic (PLEG): container finished" podID="534791d6-9c49-4e40-8269-b56d5bac5221" containerID="cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913" exitCode=0 Dec 09 17:53:13 crc kubenswrapper[4954]: I1209 17:53:13.125503 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9csc9" event={"ID":"534791d6-9c49-4e40-8269-b56d5bac5221","Type":"ContainerDied","Data":"cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913"} Dec 09 17:53:13 crc kubenswrapper[4954]: I1209 17:53:13.125533 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9csc9" event={"ID":"534791d6-9c49-4e40-8269-b56d5bac5221","Type":"ContainerStarted","Data":"87f3dd5bcb966952c0a0acdc4b70f9e97dcb71b8b444c120129749ab2dfa0e80"} Dec 09 17:53:14 crc kubenswrapper[4954]: I1209 17:53:14.154073 4954 generic.go:334] "Generic (PLEG): container finished" podID="0575ea51-efad-4a28-9cf1-b2b2653312d1" containerID="a54d3639ba9f450cbc3f77a6d92266fed780e716bc36d641deaffcfae7e140a8" exitCode=2 Dec 09 17:53:14 crc kubenswrapper[4954]: I1209 17:53:14.154528 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" event={"ID":"0575ea51-efad-4a28-9cf1-b2b2653312d1","Type":"ContainerDied","Data":"a54d3639ba9f450cbc3f77a6d92266fed780e716bc36d641deaffcfae7e140a8"} Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.173778 4954 generic.go:334] "Generic (PLEG): container finished" podID="534791d6-9c49-4e40-8269-b56d5bac5221" containerID="dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5" exitCode=0 Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.174280 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9csc9" event={"ID":"534791d6-9c49-4e40-8269-b56d5bac5221","Type":"ContainerDied","Data":"dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5"} Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.685968 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.818770 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-1\") pod \"0575ea51-efad-4a28-9cf1-b2b2653312d1\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.818860 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-0\") pod \"0575ea51-efad-4a28-9cf1-b2b2653312d1\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.818986 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-2\") pod \"0575ea51-efad-4a28-9cf1-b2b2653312d1\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.819099 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-inventory\") pod \"0575ea51-efad-4a28-9cf1-b2b2653312d1\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.819135 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ssh-key\") pod \"0575ea51-efad-4a28-9cf1-b2b2653312d1\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.819207 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbcr8\" (UniqueName: \"kubernetes.io/projected/0575ea51-efad-4a28-9cf1-b2b2653312d1-kube-api-access-jbcr8\") pod \"0575ea51-efad-4a28-9cf1-b2b2653312d1\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.819225 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-telemetry-combined-ca-bundle\") pod \"0575ea51-efad-4a28-9cf1-b2b2653312d1\" (UID: \"0575ea51-efad-4a28-9cf1-b2b2653312d1\") " Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.826828 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "0575ea51-efad-4a28-9cf1-b2b2653312d1" (UID: "0575ea51-efad-4a28-9cf1-b2b2653312d1"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.850900 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0575ea51-efad-4a28-9cf1-b2b2653312d1-kube-api-access-jbcr8" (OuterVolumeSpecName: "kube-api-access-jbcr8") pod "0575ea51-efad-4a28-9cf1-b2b2653312d1" (UID: "0575ea51-efad-4a28-9cf1-b2b2653312d1"). InnerVolumeSpecName "kube-api-access-jbcr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.866537 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "0575ea51-efad-4a28-9cf1-b2b2653312d1" (UID: "0575ea51-efad-4a28-9cf1-b2b2653312d1"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.868615 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-inventory" (OuterVolumeSpecName: "inventory") pod "0575ea51-efad-4a28-9cf1-b2b2653312d1" (UID: "0575ea51-efad-4a28-9cf1-b2b2653312d1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.876568 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "0575ea51-efad-4a28-9cf1-b2b2653312d1" (UID: "0575ea51-efad-4a28-9cf1-b2b2653312d1"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.877915 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0575ea51-efad-4a28-9cf1-b2b2653312d1" (UID: "0575ea51-efad-4a28-9cf1-b2b2653312d1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.887925 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "0575ea51-efad-4a28-9cf1-b2b2653312d1" (UID: "0575ea51-efad-4a28-9cf1-b2b2653312d1"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.921876 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.921951 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.921962 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.921972 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.921982 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbcr8\" (UniqueName: \"kubernetes.io/projected/0575ea51-efad-4a28-9cf1-b2b2653312d1-kube-api-access-jbcr8\") on node \"crc\" DevicePath \"\"" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.921991 4954 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:53:15 crc kubenswrapper[4954]: I1209 17:53:15.921999 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0575ea51-efad-4a28-9cf1-b2b2653312d1-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 09 17:53:16 crc kubenswrapper[4954]: I1209 17:53:16.120050 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:53:16 crc kubenswrapper[4954]: E1209 17:53:16.120594 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:53:16 crc kubenswrapper[4954]: I1209 17:53:16.186047 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9csc9" event={"ID":"534791d6-9c49-4e40-8269-b56d5bac5221","Type":"ContainerStarted","Data":"6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd"} Dec 09 17:53:16 crc kubenswrapper[4954]: I1209 17:53:16.188511 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" event={"ID":"0575ea51-efad-4a28-9cf1-b2b2653312d1","Type":"ContainerDied","Data":"752fefb1eadf8fd27c67518098a268d77d2f48f56171cb417f60c2852ab47be0"} Dec 09 17:53:16 crc kubenswrapper[4954]: I1209 17:53:16.188545 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="752fefb1eadf8fd27c67518098a268d77d2f48f56171cb417f60c2852ab47be0" Dec 09 17:53:16 crc kubenswrapper[4954]: I1209 17:53:16.188623 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl" Dec 09 17:53:16 crc kubenswrapper[4954]: I1209 17:53:16.218943 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9csc9" podStartSLOduration=2.532029623 podStartE2EDuration="5.218922523s" podCreationTimestamp="2025-12-09 17:53:11 +0000 UTC" firstStartedPulling="2025-12-09 17:53:13.127797974 +0000 UTC m=+3389.515971794" lastFinishedPulling="2025-12-09 17:53:15.814690874 +0000 UTC m=+3392.202864694" observedRunningTime="2025-12-09 17:53:16.209597698 +0000 UTC m=+3392.597771528" watchObservedRunningTime="2025-12-09 17:53:16.218922523 +0000 UTC m=+3392.607096343" Dec 09 17:53:17 crc kubenswrapper[4954]: E1209 17:53:17.122904 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:53:21 crc kubenswrapper[4954]: I1209 17:53:21.663737 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:21 crc kubenswrapper[4954]: I1209 17:53:21.664506 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:21 crc kubenswrapper[4954]: I1209 17:53:21.733264 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:22 crc kubenswrapper[4954]: I1209 17:53:22.294615 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:23 crc kubenswrapper[4954]: E1209 17:53:23.122891 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:53:25 crc kubenswrapper[4954]: I1209 17:53:25.327784 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9csc9"] Dec 09 17:53:25 crc kubenswrapper[4954]: I1209 17:53:25.328398 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9csc9" podUID="534791d6-9c49-4e40-8269-b56d5bac5221" containerName="registry-server" containerID="cri-o://6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd" gracePeriod=2 Dec 09 17:53:25 crc kubenswrapper[4954]: I1209 17:53:25.862877 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:25 crc kubenswrapper[4954]: I1209 17:53:25.946187 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-utilities\") pod \"534791d6-9c49-4e40-8269-b56d5bac5221\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " Dec 09 17:53:25 crc kubenswrapper[4954]: I1209 17:53:25.946294 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn6hk\" (UniqueName: \"kubernetes.io/projected/534791d6-9c49-4e40-8269-b56d5bac5221-kube-api-access-qn6hk\") pod \"534791d6-9c49-4e40-8269-b56d5bac5221\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " Dec 09 17:53:25 crc kubenswrapper[4954]: I1209 17:53:25.946430 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-catalog-content\") pod \"534791d6-9c49-4e40-8269-b56d5bac5221\" (UID: \"534791d6-9c49-4e40-8269-b56d5bac5221\") " Dec 09 17:53:25 crc kubenswrapper[4954]: I1209 17:53:25.947670 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-utilities" (OuterVolumeSpecName: "utilities") pod "534791d6-9c49-4e40-8269-b56d5bac5221" (UID: "534791d6-9c49-4e40-8269-b56d5bac5221"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:53:25 crc kubenswrapper[4954]: I1209 17:53:25.952304 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/534791d6-9c49-4e40-8269-b56d5bac5221-kube-api-access-qn6hk" (OuterVolumeSpecName: "kube-api-access-qn6hk") pod "534791d6-9c49-4e40-8269-b56d5bac5221" (UID: "534791d6-9c49-4e40-8269-b56d5bac5221"). InnerVolumeSpecName "kube-api-access-qn6hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:53:25 crc kubenswrapper[4954]: I1209 17:53:25.967381 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "534791d6-9c49-4e40-8269-b56d5bac5221" (UID: "534791d6-9c49-4e40-8269-b56d5bac5221"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.048917 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.048955 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534791d6-9c49-4e40-8269-b56d5bac5221-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.048966 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn6hk\" (UniqueName: \"kubernetes.io/projected/534791d6-9c49-4e40-8269-b56d5bac5221-kube-api-access-qn6hk\") on node \"crc\" DevicePath \"\"" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.294708 4954 generic.go:334] "Generic (PLEG): container finished" podID="534791d6-9c49-4e40-8269-b56d5bac5221" containerID="6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd" exitCode=0 Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.294758 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9csc9" event={"ID":"534791d6-9c49-4e40-8269-b56d5bac5221","Type":"ContainerDied","Data":"6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd"} Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.294787 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9csc9" event={"ID":"534791d6-9c49-4e40-8269-b56d5bac5221","Type":"ContainerDied","Data":"87f3dd5bcb966952c0a0acdc4b70f9e97dcb71b8b444c120129749ab2dfa0e80"} Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.294808 4954 scope.go:117] "RemoveContainer" containerID="6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.294964 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9csc9" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.323070 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9csc9"] Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.327639 4954 scope.go:117] "RemoveContainer" containerID="dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.337435 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9csc9"] Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.353424 4954 scope.go:117] "RemoveContainer" containerID="cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.406443 4954 scope.go:117] "RemoveContainer" containerID="6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd" Dec 09 17:53:26 crc kubenswrapper[4954]: E1209 17:53:26.406770 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd\": container with ID starting with 6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd not found: ID does not exist" containerID="6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.406802 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd"} err="failed to get container status \"6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd\": rpc error: code = NotFound desc = could not find container \"6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd\": container with ID starting with 6a7d1ef7067e727b2b78eabb8d0b08124a945f1faf18ed638933b8aa01bbf3bd not found: ID does not exist" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.406824 4954 scope.go:117] "RemoveContainer" containerID="dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5" Dec 09 17:53:26 crc kubenswrapper[4954]: E1209 17:53:26.407239 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5\": container with ID starting with dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5 not found: ID does not exist" containerID="dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.407265 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5"} err="failed to get container status \"dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5\": rpc error: code = NotFound desc = could not find container \"dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5\": container with ID starting with dcd47d57d77c7026179d4117be738f9302163c58dc0fd9092fbeff7725ef91a5 not found: ID does not exist" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.407282 4954 scope.go:117] "RemoveContainer" containerID="cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913" Dec 09 17:53:26 crc kubenswrapper[4954]: E1209 17:53:26.407607 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913\": container with ID starting with cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913 not found: ID does not exist" containerID="cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913" Dec 09 17:53:26 crc kubenswrapper[4954]: I1209 17:53:26.407659 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913"} err="failed to get container status \"cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913\": rpc error: code = NotFound desc = could not find container \"cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913\": container with ID starting with cb33d5f661a62e66e586480ff976f09cc040e0ea5fc3399bff7f80f576ea6913 not found: ID does not exist" Dec 09 17:53:28 crc kubenswrapper[4954]: I1209 17:53:28.134632 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="534791d6-9c49-4e40-8269-b56d5bac5221" path="/var/lib/kubelet/pods/534791d6-9c49-4e40-8269-b56d5bac5221/volumes" Dec 09 17:53:29 crc kubenswrapper[4954]: I1209 17:53:29.119645 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:53:29 crc kubenswrapper[4954]: E1209 17:53:29.120277 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:53:31 crc kubenswrapper[4954]: E1209 17:53:31.123647 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:53:34 crc kubenswrapper[4954]: E1209 17:53:34.128396 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:53:43 crc kubenswrapper[4954]: I1209 17:53:43.119931 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:53:43 crc kubenswrapper[4954]: E1209 17:53:43.120916 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 17:53:46 crc kubenswrapper[4954]: E1209 17:53:46.122348 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:53:48 crc kubenswrapper[4954]: E1209 17:53:48.127621 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:53:58 crc kubenswrapper[4954]: I1209 17:53:58.121806 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:53:58 crc kubenswrapper[4954]: I1209 17:53:58.600131 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"fe0720c00490598321beec492afb6905a7531ef531ec84ef61cc45bd27f3ed72"} Dec 09 17:53:59 crc kubenswrapper[4954]: E1209 17:53:59.125119 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:54:02 crc kubenswrapper[4954]: E1209 17:54:02.123968 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:54:13 crc kubenswrapper[4954]: E1209 17:54:13.135019 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:54:14 crc kubenswrapper[4954]: E1209 17:54:14.131577 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.393618 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hzm9s"] Dec 09 17:54:18 crc kubenswrapper[4954]: E1209 17:54:18.396907 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0575ea51-efad-4a28-9cf1-b2b2653312d1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.396944 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="0575ea51-efad-4a28-9cf1-b2b2653312d1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:54:18 crc kubenswrapper[4954]: E1209 17:54:18.396976 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534791d6-9c49-4e40-8269-b56d5bac5221" containerName="extract-content" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.396985 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="534791d6-9c49-4e40-8269-b56d5bac5221" containerName="extract-content" Dec 09 17:54:18 crc kubenswrapper[4954]: E1209 17:54:18.397043 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534791d6-9c49-4e40-8269-b56d5bac5221" containerName="extract-utilities" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.397052 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="534791d6-9c49-4e40-8269-b56d5bac5221" containerName="extract-utilities" Dec 09 17:54:18 crc kubenswrapper[4954]: E1209 17:54:18.397069 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534791d6-9c49-4e40-8269-b56d5bac5221" containerName="registry-server" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.397077 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="534791d6-9c49-4e40-8269-b56d5bac5221" containerName="registry-server" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.397433 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="0575ea51-efad-4a28-9cf1-b2b2653312d1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.397467 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="534791d6-9c49-4e40-8269-b56d5bac5221" containerName="registry-server" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.399312 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.413499 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hzm9s"] Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.434269 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-utilities\") pod \"redhat-operators-hzm9s\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.434367 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-catalog-content\") pod \"redhat-operators-hzm9s\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.434443 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgdnd\" (UniqueName: \"kubernetes.io/projected/377097ee-e4ed-4f9f-849a-0e04ad072aed-kube-api-access-kgdnd\") pod \"redhat-operators-hzm9s\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.538087 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgdnd\" (UniqueName: \"kubernetes.io/projected/377097ee-e4ed-4f9f-849a-0e04ad072aed-kube-api-access-kgdnd\") pod \"redhat-operators-hzm9s\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.538768 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-utilities\") pod \"redhat-operators-hzm9s\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.538887 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-catalog-content\") pod \"redhat-operators-hzm9s\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.539266 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-utilities\") pod \"redhat-operators-hzm9s\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.539362 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-catalog-content\") pod \"redhat-operators-hzm9s\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.559395 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgdnd\" (UniqueName: \"kubernetes.io/projected/377097ee-e4ed-4f9f-849a-0e04ad072aed-kube-api-access-kgdnd\") pod \"redhat-operators-hzm9s\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:18 crc kubenswrapper[4954]: I1209 17:54:18.734212 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:19 crc kubenswrapper[4954]: I1209 17:54:19.295313 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hzm9s"] Dec 09 17:54:19 crc kubenswrapper[4954]: I1209 17:54:19.800886 4954 generic.go:334] "Generic (PLEG): container finished" podID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerID="3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76" exitCode=0 Dec 09 17:54:19 crc kubenswrapper[4954]: I1209 17:54:19.800943 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzm9s" event={"ID":"377097ee-e4ed-4f9f-849a-0e04ad072aed","Type":"ContainerDied","Data":"3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76"} Dec 09 17:54:19 crc kubenswrapper[4954]: I1209 17:54:19.800975 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzm9s" event={"ID":"377097ee-e4ed-4f9f-849a-0e04ad072aed","Type":"ContainerStarted","Data":"4e74b79a977d1a5ff2e5f74c0f3e6dcce4d2c0c67a1f070b8ce6630afeacbb69"} Dec 09 17:54:20 crc kubenswrapper[4954]: I1209 17:54:20.828405 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzm9s" event={"ID":"377097ee-e4ed-4f9f-849a-0e04ad072aed","Type":"ContainerStarted","Data":"9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29"} Dec 09 17:54:23 crc kubenswrapper[4954]: I1209 17:54:23.866919 4954 generic.go:334] "Generic (PLEG): container finished" podID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerID="9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29" exitCode=0 Dec 09 17:54:23 crc kubenswrapper[4954]: I1209 17:54:23.866997 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzm9s" event={"ID":"377097ee-e4ed-4f9f-849a-0e04ad072aed","Type":"ContainerDied","Data":"9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29"} Dec 09 17:54:25 crc kubenswrapper[4954]: E1209 17:54:25.121984 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:54:25 crc kubenswrapper[4954]: I1209 17:54:25.887315 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzm9s" event={"ID":"377097ee-e4ed-4f9f-849a-0e04ad072aed","Type":"ContainerStarted","Data":"51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba"} Dec 09 17:54:25 crc kubenswrapper[4954]: I1209 17:54:25.911478 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hzm9s" podStartSLOduration=2.548002502 podStartE2EDuration="7.911455797s" podCreationTimestamp="2025-12-09 17:54:18 +0000 UTC" firstStartedPulling="2025-12-09 17:54:19.802575832 +0000 UTC m=+3456.190749662" lastFinishedPulling="2025-12-09 17:54:25.166029127 +0000 UTC m=+3461.554202957" observedRunningTime="2025-12-09 17:54:25.906665295 +0000 UTC m=+3462.294839115" watchObservedRunningTime="2025-12-09 17:54:25.911455797 +0000 UTC m=+3462.299629617" Dec 09 17:54:27 crc kubenswrapper[4954]: E1209 17:54:27.123147 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:54:28 crc kubenswrapper[4954]: I1209 17:54:28.735896 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:28 crc kubenswrapper[4954]: I1209 17:54:28.736488 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:29 crc kubenswrapper[4954]: I1209 17:54:29.821742 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hzm9s" podUID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerName="registry-server" probeResult="failure" output=< Dec 09 17:54:29 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 17:54:29 crc kubenswrapper[4954]: > Dec 09 17:54:38 crc kubenswrapper[4954]: I1209 17:54:38.785398 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:38 crc kubenswrapper[4954]: I1209 17:54:38.839616 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:39 crc kubenswrapper[4954]: I1209 17:54:39.029365 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hzm9s"] Dec 09 17:54:39 crc kubenswrapper[4954]: E1209 17:54:39.121927 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.023709 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hzm9s" podUID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerName="registry-server" containerID="cri-o://51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba" gracePeriod=2 Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.521542 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.633947 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-utilities\") pod \"377097ee-e4ed-4f9f-849a-0e04ad072aed\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.634048 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgdnd\" (UniqueName: \"kubernetes.io/projected/377097ee-e4ed-4f9f-849a-0e04ad072aed-kube-api-access-kgdnd\") pod \"377097ee-e4ed-4f9f-849a-0e04ad072aed\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.634323 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-catalog-content\") pod \"377097ee-e4ed-4f9f-849a-0e04ad072aed\" (UID: \"377097ee-e4ed-4f9f-849a-0e04ad072aed\") " Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.634724 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-utilities" (OuterVolumeSpecName: "utilities") pod "377097ee-e4ed-4f9f-849a-0e04ad072aed" (UID: "377097ee-e4ed-4f9f-849a-0e04ad072aed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.634991 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.645768 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/377097ee-e4ed-4f9f-849a-0e04ad072aed-kube-api-access-kgdnd" (OuterVolumeSpecName: "kube-api-access-kgdnd") pod "377097ee-e4ed-4f9f-849a-0e04ad072aed" (UID: "377097ee-e4ed-4f9f-849a-0e04ad072aed"). InnerVolumeSpecName "kube-api-access-kgdnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.736664 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgdnd\" (UniqueName: \"kubernetes.io/projected/377097ee-e4ed-4f9f-849a-0e04ad072aed-kube-api-access-kgdnd\") on node \"crc\" DevicePath \"\"" Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.752427 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "377097ee-e4ed-4f9f-849a-0e04ad072aed" (UID: "377097ee-e4ed-4f9f-849a-0e04ad072aed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:54:40 crc kubenswrapper[4954]: I1209 17:54:40.838180 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/377097ee-e4ed-4f9f-849a-0e04ad072aed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.038640 4954 generic.go:334] "Generic (PLEG): container finished" podID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerID="51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba" exitCode=0 Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.038697 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzm9s" event={"ID":"377097ee-e4ed-4f9f-849a-0e04ad072aed","Type":"ContainerDied","Data":"51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba"} Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.038763 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hzm9s" event={"ID":"377097ee-e4ed-4f9f-849a-0e04ad072aed","Type":"ContainerDied","Data":"4e74b79a977d1a5ff2e5f74c0f3e6dcce4d2c0c67a1f070b8ce6630afeacbb69"} Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.038806 4954 scope.go:117] "RemoveContainer" containerID="51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba" Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.038708 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hzm9s" Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.069302 4954 scope.go:117] "RemoveContainer" containerID="9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29" Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.074283 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hzm9s"] Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.086292 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hzm9s"] Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.101691 4954 scope.go:117] "RemoveContainer" containerID="3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76" Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.154548 4954 scope.go:117] "RemoveContainer" containerID="51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba" Dec 09 17:54:41 crc kubenswrapper[4954]: E1209 17:54:41.154932 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba\": container with ID starting with 51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba not found: ID does not exist" containerID="51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba" Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.154973 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba"} err="failed to get container status \"51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba\": rpc error: code = NotFound desc = could not find container \"51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba\": container with ID starting with 51689aca2eaf3e321f3158678b11cb092fa562ea793fb35f2e5665a2f91090ba not found: ID does not exist" Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.155001 4954 scope.go:117] "RemoveContainer" containerID="9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29" Dec 09 17:54:41 crc kubenswrapper[4954]: E1209 17:54:41.155294 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29\": container with ID starting with 9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29 not found: ID does not exist" containerID="9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29" Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.155347 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29"} err="failed to get container status \"9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29\": rpc error: code = NotFound desc = could not find container \"9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29\": container with ID starting with 9018e43e1849e1af164642317b15be27fb3d5b655620ae1812e6c68e38833c29 not found: ID does not exist" Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.155383 4954 scope.go:117] "RemoveContainer" containerID="3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76" Dec 09 17:54:41 crc kubenswrapper[4954]: E1209 17:54:41.155813 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76\": container with ID starting with 3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76 not found: ID does not exist" containerID="3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76" Dec 09 17:54:41 crc kubenswrapper[4954]: I1209 17:54:41.155843 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76"} err="failed to get container status \"3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76\": rpc error: code = NotFound desc = could not find container \"3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76\": container with ID starting with 3b1cd70d496cc2aec131598f2c1a3ccd9037ff0178c22a577de3712598725a76 not found: ID does not exist" Dec 09 17:54:42 crc kubenswrapper[4954]: E1209 17:54:42.123073 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:54:42 crc kubenswrapper[4954]: I1209 17:54:42.134366 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="377097ee-e4ed-4f9f-849a-0e04ad072aed" path="/var/lib/kubelet/pods/377097ee-e4ed-4f9f-849a-0e04ad072aed/volumes" Dec 09 17:54:52 crc kubenswrapper[4954]: E1209 17:54:52.123351 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:54:53 crc kubenswrapper[4954]: E1209 17:54:53.121967 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:55:03 crc kubenswrapper[4954]: E1209 17:55:03.128124 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:55:05 crc kubenswrapper[4954]: E1209 17:55:05.122790 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:55:15 crc kubenswrapper[4954]: E1209 17:55:15.122243 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:55:17 crc kubenswrapper[4954]: E1209 17:55:17.123194 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:55:27 crc kubenswrapper[4954]: E1209 17:55:27.123994 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:55:29 crc kubenswrapper[4954]: E1209 17:55:29.122257 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:55:42 crc kubenswrapper[4954]: E1209 17:55:42.122788 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:55:42 crc kubenswrapper[4954]: E1209 17:55:42.122804 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.046049 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm"] Dec 09 17:55:53 crc kubenswrapper[4954]: E1209 17:55:53.048051 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerName="registry-server" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.048077 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerName="registry-server" Dec 09 17:55:53 crc kubenswrapper[4954]: E1209 17:55:53.048108 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerName="extract-utilities" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.048115 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerName="extract-utilities" Dec 09 17:55:53 crc kubenswrapper[4954]: E1209 17:55:53.048131 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerName="extract-content" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.048139 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerName="extract-content" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.048354 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="377097ee-e4ed-4f9f-849a-0e04ad072aed" containerName="registry-server" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.049151 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.052425 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.052425 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.052582 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.052652 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.052746 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.056794 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm"] Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.119751 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.119814 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9psk8\" (UniqueName: \"kubernetes.io/projected/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-kube-api-access-9psk8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.119880 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.119916 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.120461 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.120557 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.120585 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.226144 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.226240 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9psk8\" (UniqueName: \"kubernetes.io/projected/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-kube-api-access-9psk8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.226289 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.226339 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.226477 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.226634 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.226670 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.234792 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.235280 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.235385 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.236893 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.237514 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.251310 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.253529 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9psk8\" (UniqueName: \"kubernetes.io/projected/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-kube-api-access-9psk8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.372320 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:55:53 crc kubenswrapper[4954]: I1209 17:55:53.905010 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm"] Dec 09 17:55:54 crc kubenswrapper[4954]: E1209 17:55:54.130271 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:55:54 crc kubenswrapper[4954]: I1209 17:55:54.773456 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" event={"ID":"5aace7ec-d1b0-4c50-890a-9c83e3e530fc","Type":"ContainerStarted","Data":"87776e69eaa24e2953cbb0394eabc509a5b7884b3107d3a7dc95ec685582cee5"} Dec 09 17:55:55 crc kubenswrapper[4954]: E1209 17:55:55.122886 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:55:55 crc kubenswrapper[4954]: I1209 17:55:55.787930 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" event={"ID":"5aace7ec-d1b0-4c50-890a-9c83e3e530fc","Type":"ContainerStarted","Data":"832c642c782de3270488097b9359614a2922ae45921661ddaafd0540281dfd17"} Dec 09 17:55:55 crc kubenswrapper[4954]: I1209 17:55:55.826143 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" podStartSLOduration=2.157352792 podStartE2EDuration="2.826112594s" podCreationTimestamp="2025-12-09 17:55:53 +0000 UTC" firstStartedPulling="2025-12-09 17:55:53.9089688 +0000 UTC m=+3550.297142620" lastFinishedPulling="2025-12-09 17:55:54.577728602 +0000 UTC m=+3550.965902422" observedRunningTime="2025-12-09 17:55:55.818560306 +0000 UTC m=+3552.206734136" watchObservedRunningTime="2025-12-09 17:55:55.826112594 +0000 UTC m=+3552.214286454" Dec 09 17:56:05 crc kubenswrapper[4954]: E1209 17:56:05.124332 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:56:06 crc kubenswrapper[4954]: E1209 17:56:06.122479 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:56:13 crc kubenswrapper[4954]: I1209 17:56:13.755215 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:56:13 crc kubenswrapper[4954]: I1209 17:56:13.755812 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:56:17 crc kubenswrapper[4954]: E1209 17:56:17.123723 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:56:20 crc kubenswrapper[4954]: E1209 17:56:20.123849 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:56:31 crc kubenswrapper[4954]: E1209 17:56:31.256797 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:56:35 crc kubenswrapper[4954]: E1209 17:56:35.122641 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:56:43 crc kubenswrapper[4954]: E1209 17:56:43.123613 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:56:43 crc kubenswrapper[4954]: I1209 17:56:43.755717 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:56:43 crc kubenswrapper[4954]: I1209 17:56:43.756259 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:56:46 crc kubenswrapper[4954]: E1209 17:56:46.124300 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:56:47 crc kubenswrapper[4954]: I1209 17:56:47.500189 4954 generic.go:334] "Generic (PLEG): container finished" podID="5aace7ec-d1b0-4c50-890a-9c83e3e530fc" containerID="832c642c782de3270488097b9359614a2922ae45921661ddaafd0540281dfd17" exitCode=2 Dec 09 17:56:47 crc kubenswrapper[4954]: I1209 17:56:47.500541 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" event={"ID":"5aace7ec-d1b0-4c50-890a-9c83e3e530fc","Type":"ContainerDied","Data":"832c642c782de3270488097b9359614a2922ae45921661ddaafd0540281dfd17"} Dec 09 17:56:48 crc kubenswrapper[4954]: I1209 17:56:48.936657 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.081960 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-0\") pod \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.082412 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-inventory\") pod \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.082478 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9psk8\" (UniqueName: \"kubernetes.io/projected/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-kube-api-access-9psk8\") pod \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.082623 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-2\") pod \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.082682 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ssh-key\") pod \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.082872 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-telemetry-combined-ca-bundle\") pod \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.082936 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-1\") pod \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\" (UID: \"5aace7ec-d1b0-4c50-890a-9c83e3e530fc\") " Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.088495 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "5aace7ec-d1b0-4c50-890a-9c83e3e530fc" (UID: "5aace7ec-d1b0-4c50-890a-9c83e3e530fc"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.088810 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-kube-api-access-9psk8" (OuterVolumeSpecName: "kube-api-access-9psk8") pod "5aace7ec-d1b0-4c50-890a-9c83e3e530fc" (UID: "5aace7ec-d1b0-4c50-890a-9c83e3e530fc"). InnerVolumeSpecName "kube-api-access-9psk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.116152 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5aace7ec-d1b0-4c50-890a-9c83e3e530fc" (UID: "5aace7ec-d1b0-4c50-890a-9c83e3e530fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.118262 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "5aace7ec-d1b0-4c50-890a-9c83e3e530fc" (UID: "5aace7ec-d1b0-4c50-890a-9c83e3e530fc"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.119616 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "5aace7ec-d1b0-4c50-890a-9c83e3e530fc" (UID: "5aace7ec-d1b0-4c50-890a-9c83e3e530fc"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.121986 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "5aace7ec-d1b0-4c50-890a-9c83e3e530fc" (UID: "5aace7ec-d1b0-4c50-890a-9c83e3e530fc"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.134852 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-inventory" (OuterVolumeSpecName: "inventory") pod "5aace7ec-d1b0-4c50-890a-9c83e3e530fc" (UID: "5aace7ec-d1b0-4c50-890a-9c83e3e530fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.185453 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9psk8\" (UniqueName: \"kubernetes.io/projected/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-kube-api-access-9psk8\") on node \"crc\" DevicePath \"\"" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.185488 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.185503 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.185518 4954 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.185532 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.185544 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.185559 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5aace7ec-d1b0-4c50-890a-9c83e3e530fc-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.519908 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" event={"ID":"5aace7ec-d1b0-4c50-890a-9c83e3e530fc","Type":"ContainerDied","Data":"87776e69eaa24e2953cbb0394eabc509a5b7884b3107d3a7dc95ec685582cee5"} Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.519953 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87776e69eaa24e2953cbb0394eabc509a5b7884b3107d3a7dc95ec685582cee5" Dec 09 17:56:49 crc kubenswrapper[4954]: I1209 17:56:49.519983 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm" Dec 09 17:56:58 crc kubenswrapper[4954]: E1209 17:56:58.123529 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:56:59 crc kubenswrapper[4954]: E1209 17:56:59.122538 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:57:05 crc kubenswrapper[4954]: I1209 17:57:05.425923 4954 scope.go:117] "RemoveContainer" containerID="f2115e5c7a81ff6b6c6ac882e03bb015308783388af6a6b81834a906d9051618" Dec 09 17:57:05 crc kubenswrapper[4954]: I1209 17:57:05.454812 4954 scope.go:117] "RemoveContainer" containerID="689e8eae7865a47cbab6a244c5f8b2bb5e5806a20c53ff7fe81d377c22c8f412" Dec 09 17:57:05 crc kubenswrapper[4954]: I1209 17:57:05.544722 4954 scope.go:117] "RemoveContainer" containerID="4cc9f45c0cd874169c88aae1965894bb44264fa653f2fe7d0828debbd697e39e" Dec 09 17:57:10 crc kubenswrapper[4954]: E1209 17:57:10.122742 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:57:11 crc kubenswrapper[4954]: E1209 17:57:11.122514 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:57:13 crc kubenswrapper[4954]: I1209 17:57:13.755887 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:57:13 crc kubenswrapper[4954]: I1209 17:57:13.756487 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:57:13 crc kubenswrapper[4954]: I1209 17:57:13.756576 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 17:57:13 crc kubenswrapper[4954]: I1209 17:57:13.757997 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe0720c00490598321beec492afb6905a7531ef531ec84ef61cc45bd27f3ed72"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 17:57:13 crc kubenswrapper[4954]: I1209 17:57:13.758109 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://fe0720c00490598321beec492afb6905a7531ef531ec84ef61cc45bd27f3ed72" gracePeriod=600 Dec 09 17:57:14 crc kubenswrapper[4954]: I1209 17:57:14.877662 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="fe0720c00490598321beec492afb6905a7531ef531ec84ef61cc45bd27f3ed72" exitCode=0 Dec 09 17:57:14 crc kubenswrapper[4954]: I1209 17:57:14.877736 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"fe0720c00490598321beec492afb6905a7531ef531ec84ef61cc45bd27f3ed72"} Dec 09 17:57:14 crc kubenswrapper[4954]: I1209 17:57:14.878219 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4"} Dec 09 17:57:14 crc kubenswrapper[4954]: I1209 17:57:14.878247 4954 scope.go:117] "RemoveContainer" containerID="9f62c4934273e6e5b40e2b6d5cf3ece12f0e42b70b316c3722d40248344a970f" Dec 09 17:57:21 crc kubenswrapper[4954]: E1209 17:57:21.122714 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:57:23 crc kubenswrapper[4954]: E1209 17:57:23.123424 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:57:36 crc kubenswrapper[4954]: I1209 17:57:36.127074 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 17:57:36 crc kubenswrapper[4954]: E1209 17:57:36.288065 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:57:36 crc kubenswrapper[4954]: E1209 17:57:36.288723 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 17:57:36 crc kubenswrapper[4954]: E1209 17:57:36.289071 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:57:36 crc kubenswrapper[4954]: E1209 17:57:36.290331 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:57:37 crc kubenswrapper[4954]: E1209 17:57:37.122349 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:57:48 crc kubenswrapper[4954]: E1209 17:57:48.123572 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:57:51 crc kubenswrapper[4954]: E1209 17:57:51.123230 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:57:59 crc kubenswrapper[4954]: E1209 17:57:59.123999 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:58:02 crc kubenswrapper[4954]: E1209 17:58:02.215087 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:58:02 crc kubenswrapper[4954]: E1209 17:58:02.215682 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 17:58:02 crc kubenswrapper[4954]: E1209 17:58:02.215832 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 17:58:02 crc kubenswrapper[4954]: E1209 17:58:02.217086 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.699629 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cbsp7"] Dec 09 17:58:09 crc kubenswrapper[4954]: E1209 17:58:09.700719 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aace7ec-d1b0-4c50-890a-9c83e3e530fc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.700734 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aace7ec-d1b0-4c50-890a-9c83e3e530fc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.701746 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aace7ec-d1b0-4c50-890a-9c83e3e530fc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.708208 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.752720 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cbsp7"] Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.883388 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r2l9\" (UniqueName: \"kubernetes.io/projected/0055c556-4fda-424a-b590-5a6ca19d8314-kube-api-access-5r2l9\") pod \"certified-operators-cbsp7\" (UID: \"0055c556-4fda-424a-b590-5a6ca19d8314\") " pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.883460 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0055c556-4fda-424a-b590-5a6ca19d8314-catalog-content\") pod \"certified-operators-cbsp7\" (UID: \"0055c556-4fda-424a-b590-5a6ca19d8314\") " pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.883582 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0055c556-4fda-424a-b590-5a6ca19d8314-utilities\") pod \"certified-operators-cbsp7\" (UID: \"0055c556-4fda-424a-b590-5a6ca19d8314\") " pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.985103 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r2l9\" (UniqueName: \"kubernetes.io/projected/0055c556-4fda-424a-b590-5a6ca19d8314-kube-api-access-5r2l9\") pod \"certified-operators-cbsp7\" (UID: \"0055c556-4fda-424a-b590-5a6ca19d8314\") " pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.985164 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0055c556-4fda-424a-b590-5a6ca19d8314-catalog-content\") pod \"certified-operators-cbsp7\" (UID: \"0055c556-4fda-424a-b590-5a6ca19d8314\") " pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.985273 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0055c556-4fda-424a-b590-5a6ca19d8314-utilities\") pod \"certified-operators-cbsp7\" (UID: \"0055c556-4fda-424a-b590-5a6ca19d8314\") " pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.985761 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0055c556-4fda-424a-b590-5a6ca19d8314-utilities\") pod \"certified-operators-cbsp7\" (UID: \"0055c556-4fda-424a-b590-5a6ca19d8314\") " pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:09 crc kubenswrapper[4954]: I1209 17:58:09.985845 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0055c556-4fda-424a-b590-5a6ca19d8314-catalog-content\") pod \"certified-operators-cbsp7\" (UID: \"0055c556-4fda-424a-b590-5a6ca19d8314\") " pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:10 crc kubenswrapper[4954]: I1209 17:58:10.004164 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r2l9\" (UniqueName: \"kubernetes.io/projected/0055c556-4fda-424a-b590-5a6ca19d8314-kube-api-access-5r2l9\") pod \"certified-operators-cbsp7\" (UID: \"0055c556-4fda-424a-b590-5a6ca19d8314\") " pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:10 crc kubenswrapper[4954]: I1209 17:58:10.035143 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:10 crc kubenswrapper[4954]: E1209 17:58:10.134047 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:58:10 crc kubenswrapper[4954]: I1209 17:58:10.491416 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cbsp7"] Dec 09 17:58:11 crc kubenswrapper[4954]: I1209 17:58:11.434162 4954 generic.go:334] "Generic (PLEG): container finished" podID="0055c556-4fda-424a-b590-5a6ca19d8314" containerID="6f6b62e1389d0b1ecd6476042efdb8c2be500ecf5835093f856f3c9ffb991e30" exitCode=0 Dec 09 17:58:11 crc kubenswrapper[4954]: I1209 17:58:11.434240 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbsp7" event={"ID":"0055c556-4fda-424a-b590-5a6ca19d8314","Type":"ContainerDied","Data":"6f6b62e1389d0b1ecd6476042efdb8c2be500ecf5835093f856f3c9ffb991e30"} Dec 09 17:58:11 crc kubenswrapper[4954]: I1209 17:58:11.434530 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbsp7" event={"ID":"0055c556-4fda-424a-b590-5a6ca19d8314","Type":"ContainerStarted","Data":"70281615d239e7ca3ddb27ab428504ea69f7b1f0a2ce6b667c9101515ba5f9cc"} Dec 09 17:58:16 crc kubenswrapper[4954]: I1209 17:58:16.486453 4954 generic.go:334] "Generic (PLEG): container finished" podID="0055c556-4fda-424a-b590-5a6ca19d8314" containerID="3892ec0d29ffe0d9a051bc4a2459a5c5dc22fa947ca8b6a42629f177c0df605e" exitCode=0 Dec 09 17:58:16 crc kubenswrapper[4954]: I1209 17:58:16.486757 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbsp7" event={"ID":"0055c556-4fda-424a-b590-5a6ca19d8314","Type":"ContainerDied","Data":"3892ec0d29ffe0d9a051bc4a2459a5c5dc22fa947ca8b6a42629f177c0df605e"} Dec 09 17:58:17 crc kubenswrapper[4954]: I1209 17:58:17.503390 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cbsp7" event={"ID":"0055c556-4fda-424a-b590-5a6ca19d8314","Type":"ContainerStarted","Data":"befa1e72217ed0c72b3ee413b11e9a6b3fa421ffd832fe452c24d627a54ab523"} Dec 09 17:58:17 crc kubenswrapper[4954]: I1209 17:58:17.536011 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cbsp7" podStartSLOduration=2.743675922 podStartE2EDuration="8.535992616s" podCreationTimestamp="2025-12-09 17:58:09 +0000 UTC" firstStartedPulling="2025-12-09 17:58:11.436728356 +0000 UTC m=+3687.824902176" lastFinishedPulling="2025-12-09 17:58:17.22904505 +0000 UTC m=+3693.617218870" observedRunningTime="2025-12-09 17:58:17.52876153 +0000 UTC m=+3693.916935350" watchObservedRunningTime="2025-12-09 17:58:17.535992616 +0000 UTC m=+3693.924166436" Dec 09 17:58:18 crc kubenswrapper[4954]: E1209 17:58:18.122227 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:58:20 crc kubenswrapper[4954]: I1209 17:58:20.035319 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:20 crc kubenswrapper[4954]: I1209 17:58:20.035770 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:20 crc kubenswrapper[4954]: I1209 17:58:20.086925 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:25 crc kubenswrapper[4954]: E1209 17:58:25.123169 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.098694 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cbsp7" Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.178355 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cbsp7"] Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.291521 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t9mg8"] Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.291788 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t9mg8" podUID="98dfad17-db42-4215-add1-3d3b01a971b0" containerName="registry-server" containerID="cri-o://01ebf81616168a4141040377a1ad3c451c3b64012972858a42beed2824510efe" gracePeriod=2 Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.640494 4954 generic.go:334] "Generic (PLEG): container finished" podID="98dfad17-db42-4215-add1-3d3b01a971b0" containerID="01ebf81616168a4141040377a1ad3c451c3b64012972858a42beed2824510efe" exitCode=0 Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.640583 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9mg8" event={"ID":"98dfad17-db42-4215-add1-3d3b01a971b0","Type":"ContainerDied","Data":"01ebf81616168a4141040377a1ad3c451c3b64012972858a42beed2824510efe"} Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.824441 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.948270 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbvnn\" (UniqueName: \"kubernetes.io/projected/98dfad17-db42-4215-add1-3d3b01a971b0-kube-api-access-bbvnn\") pod \"98dfad17-db42-4215-add1-3d3b01a971b0\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.948575 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-utilities\") pod \"98dfad17-db42-4215-add1-3d3b01a971b0\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.948778 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-catalog-content\") pod \"98dfad17-db42-4215-add1-3d3b01a971b0\" (UID: \"98dfad17-db42-4215-add1-3d3b01a971b0\") " Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.953236 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-utilities" (OuterVolumeSpecName: "utilities") pod "98dfad17-db42-4215-add1-3d3b01a971b0" (UID: "98dfad17-db42-4215-add1-3d3b01a971b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:58:30 crc kubenswrapper[4954]: I1209 17:58:30.956090 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98dfad17-db42-4215-add1-3d3b01a971b0-kube-api-access-bbvnn" (OuterVolumeSpecName: "kube-api-access-bbvnn") pod "98dfad17-db42-4215-add1-3d3b01a971b0" (UID: "98dfad17-db42-4215-add1-3d3b01a971b0"). InnerVolumeSpecName "kube-api-access-bbvnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.035609 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98dfad17-db42-4215-add1-3d3b01a971b0" (UID: "98dfad17-db42-4215-add1-3d3b01a971b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.050870 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.050925 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbvnn\" (UniqueName: \"kubernetes.io/projected/98dfad17-db42-4215-add1-3d3b01a971b0-kube-api-access-bbvnn\") on node \"crc\" DevicePath \"\"" Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.050939 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98dfad17-db42-4215-add1-3d3b01a971b0-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.651444 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9mg8" event={"ID":"98dfad17-db42-4215-add1-3d3b01a971b0","Type":"ContainerDied","Data":"9443228bff0f5a04c931e3b14c888d10f09913c7edafc3856936ebf25e853bc6"} Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.651846 4954 scope.go:117] "RemoveContainer" containerID="01ebf81616168a4141040377a1ad3c451c3b64012972858a42beed2824510efe" Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.651585 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9mg8" Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.681311 4954 scope.go:117] "RemoveContainer" containerID="a01d965c56720dbf5f22bd97da4332e2aca6463852e930ea0d400c7034578267" Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.686672 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t9mg8"] Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.696890 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t9mg8"] Dec 09 17:58:31 crc kubenswrapper[4954]: I1209 17:58:31.709758 4954 scope.go:117] "RemoveContainer" containerID="014be430bee464ff6819affc7cf16f33194c0a2572a3c4cf9ed0f0c1f69f2354" Dec 09 17:58:32 crc kubenswrapper[4954]: I1209 17:58:32.134079 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98dfad17-db42-4215-add1-3d3b01a971b0" path="/var/lib/kubelet/pods/98dfad17-db42-4215-add1-3d3b01a971b0/volumes" Dec 09 17:58:33 crc kubenswrapper[4954]: E1209 17:58:33.122102 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:58:37 crc kubenswrapper[4954]: E1209 17:58:37.122891 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:58:47 crc kubenswrapper[4954]: E1209 17:58:47.122070 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:58:50 crc kubenswrapper[4954]: E1209 17:58:50.122611 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:59:02 crc kubenswrapper[4954]: E1209 17:59:02.122366 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:59:03 crc kubenswrapper[4954]: E1209 17:59:03.124270 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:59:14 crc kubenswrapper[4954]: E1209 17:59:14.135817 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:59:15 crc kubenswrapper[4954]: E1209 17:59:15.121301 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:59:26 crc kubenswrapper[4954]: E1209 17:59:26.124451 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:59:30 crc kubenswrapper[4954]: E1209 17:59:30.122840 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:59:37 crc kubenswrapper[4954]: E1209 17:59:37.122639 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:59:41 crc kubenswrapper[4954]: E1209 17:59:41.122115 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 17:59:43 crc kubenswrapper[4954]: I1209 17:59:43.756036 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 17:59:43 crc kubenswrapper[4954]: I1209 17:59:43.756664 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 17:59:51 crc kubenswrapper[4954]: E1209 17:59:51.122752 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 17:59:55 crc kubenswrapper[4954]: E1209 17:59:55.123765 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.211073 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm"] Dec 09 18:00:00 crc kubenswrapper[4954]: E1209 18:00:00.212286 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98dfad17-db42-4215-add1-3d3b01a971b0" containerName="registry-server" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.212306 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="98dfad17-db42-4215-add1-3d3b01a971b0" containerName="registry-server" Dec 09 18:00:00 crc kubenswrapper[4954]: E1209 18:00:00.212340 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98dfad17-db42-4215-add1-3d3b01a971b0" containerName="extract-content" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.212349 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="98dfad17-db42-4215-add1-3d3b01a971b0" containerName="extract-content" Dec 09 18:00:00 crc kubenswrapper[4954]: E1209 18:00:00.212379 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98dfad17-db42-4215-add1-3d3b01a971b0" containerName="extract-utilities" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.212388 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="98dfad17-db42-4215-add1-3d3b01a971b0" containerName="extract-utilities" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.212822 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="98dfad17-db42-4215-add1-3d3b01a971b0" containerName="registry-server" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.214108 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.219777 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.219826 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.229039 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm"] Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.294215 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdr5t\" (UniqueName: \"kubernetes.io/projected/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-kube-api-access-kdr5t\") pod \"collect-profiles-29421720-hkbmm\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.294366 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-config-volume\") pod \"collect-profiles-29421720-hkbmm\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.294461 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-secret-volume\") pod \"collect-profiles-29421720-hkbmm\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.397575 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdr5t\" (UniqueName: \"kubernetes.io/projected/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-kube-api-access-kdr5t\") pod \"collect-profiles-29421720-hkbmm\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.397740 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-config-volume\") pod \"collect-profiles-29421720-hkbmm\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.397838 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-secret-volume\") pod \"collect-profiles-29421720-hkbmm\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.399971 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-config-volume\") pod \"collect-profiles-29421720-hkbmm\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.406863 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-secret-volume\") pod \"collect-profiles-29421720-hkbmm\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.423495 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdr5t\" (UniqueName: \"kubernetes.io/projected/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-kube-api-access-kdr5t\") pod \"collect-profiles-29421720-hkbmm\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:00 crc kubenswrapper[4954]: I1209 18:00:00.540003 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:01 crc kubenswrapper[4954]: W1209 18:00:01.033072 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2eea4357_2ca2_4d47_b67c_d6fc544d0ce1.slice/crio-89873cc38dc80e34d08c4b746f74550fce161519015663b71132653d2b140632 WatchSource:0}: Error finding container 89873cc38dc80e34d08c4b746f74550fce161519015663b71132653d2b140632: Status 404 returned error can't find the container with id 89873cc38dc80e34d08c4b746f74550fce161519015663b71132653d2b140632 Dec 09 18:00:01 crc kubenswrapper[4954]: I1209 18:00:01.034213 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm"] Dec 09 18:00:01 crc kubenswrapper[4954]: I1209 18:00:01.486414 4954 generic.go:334] "Generic (PLEG): container finished" podID="2eea4357-2ca2-4d47-b67c-d6fc544d0ce1" containerID="c3257ebcdf9849768733cd314a1aa62d9a8611b9927fdbd795b5f1d665b83e40" exitCode=0 Dec 09 18:00:01 crc kubenswrapper[4954]: I1209 18:00:01.486468 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" event={"ID":"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1","Type":"ContainerDied","Data":"c3257ebcdf9849768733cd314a1aa62d9a8611b9927fdbd795b5f1d665b83e40"} Dec 09 18:00:01 crc kubenswrapper[4954]: I1209 18:00:01.486499 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" event={"ID":"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1","Type":"ContainerStarted","Data":"89873cc38dc80e34d08c4b746f74550fce161519015663b71132653d2b140632"} Dec 09 18:00:02 crc kubenswrapper[4954]: I1209 18:00:02.882428 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:02 crc kubenswrapper[4954]: I1209 18:00:02.964066 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdr5t\" (UniqueName: \"kubernetes.io/projected/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-kube-api-access-kdr5t\") pod \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " Dec 09 18:00:02 crc kubenswrapper[4954]: I1209 18:00:02.964285 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-secret-volume\") pod \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " Dec 09 18:00:02 crc kubenswrapper[4954]: I1209 18:00:02.964550 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-config-volume\") pod \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\" (UID: \"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1\") " Dec 09 18:00:02 crc kubenswrapper[4954]: I1209 18:00:02.965880 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-config-volume" (OuterVolumeSpecName: "config-volume") pod "2eea4357-2ca2-4d47-b67c-d6fc544d0ce1" (UID: "2eea4357-2ca2-4d47-b67c-d6fc544d0ce1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 18:00:02 crc kubenswrapper[4954]: I1209 18:00:02.971288 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2eea4357-2ca2-4d47-b67c-d6fc544d0ce1" (UID: "2eea4357-2ca2-4d47-b67c-d6fc544d0ce1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:00:02 crc kubenswrapper[4954]: I1209 18:00:02.974130 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-kube-api-access-kdr5t" (OuterVolumeSpecName: "kube-api-access-kdr5t") pod "2eea4357-2ca2-4d47-b67c-d6fc544d0ce1" (UID: "2eea4357-2ca2-4d47-b67c-d6fc544d0ce1"). InnerVolumeSpecName "kube-api-access-kdr5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:00:03 crc kubenswrapper[4954]: I1209 18:00:03.067639 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdr5t\" (UniqueName: \"kubernetes.io/projected/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-kube-api-access-kdr5t\") on node \"crc\" DevicePath \"\"" Dec 09 18:00:03 crc kubenswrapper[4954]: I1209 18:00:03.067671 4954 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 18:00:03 crc kubenswrapper[4954]: I1209 18:00:03.067683 4954 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2eea4357-2ca2-4d47-b67c-d6fc544d0ce1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 18:00:03 crc kubenswrapper[4954]: I1209 18:00:03.515079 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" event={"ID":"2eea4357-2ca2-4d47-b67c-d6fc544d0ce1","Type":"ContainerDied","Data":"89873cc38dc80e34d08c4b746f74550fce161519015663b71132653d2b140632"} Dec 09 18:00:03 crc kubenswrapper[4954]: I1209 18:00:03.515119 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89873cc38dc80e34d08c4b746f74550fce161519015663b71132653d2b140632" Dec 09 18:00:03 crc kubenswrapper[4954]: I1209 18:00:03.515175 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421720-hkbmm" Dec 09 18:00:03 crc kubenswrapper[4954]: E1209 18:00:03.597423 4954 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2eea4357_2ca2_4d47_b67c_d6fc544d0ce1.slice\": RecentStats: unable to find data in memory cache]" Dec 09 18:00:03 crc kubenswrapper[4954]: I1209 18:00:03.959906 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5"] Dec 09 18:00:03 crc kubenswrapper[4954]: I1209 18:00:03.973490 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421675-cp6f5"] Dec 09 18:00:04 crc kubenswrapper[4954]: I1209 18:00:04.137072 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26be4df9-9c9d-47f7-b334-bbffc33b9ccb" path="/var/lib/kubelet/pods/26be4df9-9c9d-47f7-b334-bbffc33b9ccb/volumes" Dec 09 18:00:05 crc kubenswrapper[4954]: I1209 18:00:05.694995 4954 scope.go:117] "RemoveContainer" containerID="5d9cbe15162d8809d8457e6a60e98676eb62c0335eca7480f4a8746503912680" Dec 09 18:00:06 crc kubenswrapper[4954]: E1209 18:00:06.122494 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:00:08 crc kubenswrapper[4954]: E1209 18:00:08.122182 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:00:13 crc kubenswrapper[4954]: I1209 18:00:13.755157 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:00:13 crc kubenswrapper[4954]: I1209 18:00:13.755729 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:00:19 crc kubenswrapper[4954]: E1209 18:00:19.123279 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:00:23 crc kubenswrapper[4954]: E1209 18:00:23.122525 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:00:32 crc kubenswrapper[4954]: E1209 18:00:32.122292 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:00:34 crc kubenswrapper[4954]: E1209 18:00:34.131351 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:00:43 crc kubenswrapper[4954]: E1209 18:00:43.125045 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:00:43 crc kubenswrapper[4954]: I1209 18:00:43.755860 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:00:43 crc kubenswrapper[4954]: I1209 18:00:43.755928 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:00:43 crc kubenswrapper[4954]: I1209 18:00:43.755974 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 18:00:43 crc kubenswrapper[4954]: I1209 18:00:43.756841 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 18:00:43 crc kubenswrapper[4954]: I1209 18:00:43.756965 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" gracePeriod=600 Dec 09 18:00:43 crc kubenswrapper[4954]: E1209 18:00:43.876525 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:00:43 crc kubenswrapper[4954]: I1209 18:00:43.889764 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" exitCode=0 Dec 09 18:00:43 crc kubenswrapper[4954]: I1209 18:00:43.889803 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4"} Dec 09 18:00:43 crc kubenswrapper[4954]: I1209 18:00:43.889866 4954 scope.go:117] "RemoveContainer" containerID="fe0720c00490598321beec492afb6905a7531ef531ec84ef61cc45bd27f3ed72" Dec 09 18:00:43 crc kubenswrapper[4954]: I1209 18:00:43.890674 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:00:43 crc kubenswrapper[4954]: E1209 18:00:43.891008 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:00:48 crc kubenswrapper[4954]: E1209 18:00:48.122961 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:00:56 crc kubenswrapper[4954]: E1209 18:00:56.122295 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:00:59 crc kubenswrapper[4954]: I1209 18:00:59.120514 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:00:59 crc kubenswrapper[4954]: E1209 18:00:59.121380 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.206707 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29421721-b7lcm"] Dec 09 18:01:00 crc kubenswrapper[4954]: E1209 18:01:00.207383 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eea4357-2ca2-4d47-b67c-d6fc544d0ce1" containerName="collect-profiles" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.207409 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eea4357-2ca2-4d47-b67c-d6fc544d0ce1" containerName="collect-profiles" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.207760 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eea4357-2ca2-4d47-b67c-d6fc544d0ce1" containerName="collect-profiles" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.208817 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.217239 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29421721-b7lcm"] Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.382711 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-config-data\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.383170 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-combined-ca-bundle\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.383303 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmrm7\" (UniqueName: \"kubernetes.io/projected/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-kube-api-access-pmrm7\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.383427 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-fernet-keys\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.485068 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-combined-ca-bundle\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.485720 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmrm7\" (UniqueName: \"kubernetes.io/projected/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-kube-api-access-pmrm7\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.485860 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-fernet-keys\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.486087 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-config-data\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.493328 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-fernet-keys\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.493971 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-combined-ca-bundle\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.498808 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-config-data\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.504279 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmrm7\" (UniqueName: \"kubernetes.io/projected/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-kube-api-access-pmrm7\") pod \"keystone-cron-29421721-b7lcm\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.530264 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jbzhx"] Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.533186 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.544921 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jbzhx"] Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.548960 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.588723 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-catalog-content\") pod \"community-operators-jbzhx\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.588891 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fnrp\" (UniqueName: \"kubernetes.io/projected/05f88770-9339-4cd6-8aca-417259a85096-kube-api-access-5fnrp\") pod \"community-operators-jbzhx\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.588929 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-utilities\") pod \"community-operators-jbzhx\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.695144 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fnrp\" (UniqueName: \"kubernetes.io/projected/05f88770-9339-4cd6-8aca-417259a85096-kube-api-access-5fnrp\") pod \"community-operators-jbzhx\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.695457 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-utilities\") pod \"community-operators-jbzhx\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.695729 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-catalog-content\") pod \"community-operators-jbzhx\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.696718 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-catalog-content\") pod \"community-operators-jbzhx\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.696791 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-utilities\") pod \"community-operators-jbzhx\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.735790 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fnrp\" (UniqueName: \"kubernetes.io/projected/05f88770-9339-4cd6-8aca-417259a85096-kube-api-access-5fnrp\") pod \"community-operators-jbzhx\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:00 crc kubenswrapper[4954]: I1209 18:01:00.987066 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29421721-b7lcm"] Dec 09 18:01:01 crc kubenswrapper[4954]: I1209 18:01:01.016428 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:01 crc kubenswrapper[4954]: I1209 18:01:01.107419 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421721-b7lcm" event={"ID":"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d","Type":"ContainerStarted","Data":"aebcc7ebcf5da45d6ab9d914302257de05774cbbfb63f4bdbbced28742c9b314"} Dec 09 18:01:01 crc kubenswrapper[4954]: I1209 18:01:01.589109 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jbzhx"] Dec 09 18:01:02 crc kubenswrapper[4954]: I1209 18:01:02.118664 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421721-b7lcm" event={"ID":"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d","Type":"ContainerStarted","Data":"0a921c4f60833d52521a22562821231daffd3358c4736d7934c57e2d094f7d48"} Dec 09 18:01:02 crc kubenswrapper[4954]: I1209 18:01:02.122097 4954 generic.go:334] "Generic (PLEG): container finished" podID="05f88770-9339-4cd6-8aca-417259a85096" containerID="6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a" exitCode=0 Dec 09 18:01:02 crc kubenswrapper[4954]: I1209 18:01:02.137255 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzhx" event={"ID":"05f88770-9339-4cd6-8aca-417259a85096","Type":"ContainerDied","Data":"6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a"} Dec 09 18:01:02 crc kubenswrapper[4954]: I1209 18:01:02.137300 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzhx" event={"ID":"05f88770-9339-4cd6-8aca-417259a85096","Type":"ContainerStarted","Data":"43322653bc55d981676219d5929a9bd572069e90a1290ddaccd542a05d88f689"} Dec 09 18:01:02 crc kubenswrapper[4954]: I1209 18:01:02.159932 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29421721-b7lcm" podStartSLOduration=2.159896427 podStartE2EDuration="2.159896427s" podCreationTimestamp="2025-12-09 18:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 18:01:02.146134586 +0000 UTC m=+3858.534308426" watchObservedRunningTime="2025-12-09 18:01:02.159896427 +0000 UTC m=+3858.548070247" Dec 09 18:01:03 crc kubenswrapper[4954]: E1209 18:01:03.121729 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:01:04 crc kubenswrapper[4954]: I1209 18:01:04.144740 4954 generic.go:334] "Generic (PLEG): container finished" podID="05f88770-9339-4cd6-8aca-417259a85096" containerID="26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e" exitCode=0 Dec 09 18:01:04 crc kubenswrapper[4954]: I1209 18:01:04.144808 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzhx" event={"ID":"05f88770-9339-4cd6-8aca-417259a85096","Type":"ContainerDied","Data":"26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e"} Dec 09 18:01:04 crc kubenswrapper[4954]: I1209 18:01:04.149228 4954 generic.go:334] "Generic (PLEG): container finished" podID="71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d" containerID="0a921c4f60833d52521a22562821231daffd3358c4736d7934c57e2d094f7d48" exitCode=0 Dec 09 18:01:04 crc kubenswrapper[4954]: I1209 18:01:04.149295 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421721-b7lcm" event={"ID":"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d","Type":"ContainerDied","Data":"0a921c4f60833d52521a22562821231daffd3358c4736d7934c57e2d094f7d48"} Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.646182 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.718433 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmrm7\" (UniqueName: \"kubernetes.io/projected/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-kube-api-access-pmrm7\") pod \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.718501 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-config-data\") pod \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.718689 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-combined-ca-bundle\") pod \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.718756 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-fernet-keys\") pod \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\" (UID: \"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d\") " Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.724901 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d" (UID: "71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.725528 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-kube-api-access-pmrm7" (OuterVolumeSpecName: "kube-api-access-pmrm7") pod "71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d" (UID: "71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d"). InnerVolumeSpecName "kube-api-access-pmrm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.761718 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d" (UID: "71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.778855 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-config-data" (OuterVolumeSpecName: "config-data") pod "71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d" (UID: "71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.820746 4954 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.820782 4954 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.820792 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmrm7\" (UniqueName: \"kubernetes.io/projected/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-kube-api-access-pmrm7\") on node \"crc\" DevicePath \"\"" Dec 09 18:01:05 crc kubenswrapper[4954]: I1209 18:01:05.820803 4954 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 18:01:06 crc kubenswrapper[4954]: I1209 18:01:06.174609 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421721-b7lcm" event={"ID":"71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d","Type":"ContainerDied","Data":"aebcc7ebcf5da45d6ab9d914302257de05774cbbfb63f4bdbbced28742c9b314"} Dec 09 18:01:06 crc kubenswrapper[4954]: I1209 18:01:06.174674 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aebcc7ebcf5da45d6ab9d914302257de05774cbbfb63f4bdbbced28742c9b314" Dec 09 18:01:06 crc kubenswrapper[4954]: I1209 18:01:06.174625 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421721-b7lcm" Dec 09 18:01:06 crc kubenswrapper[4954]: I1209 18:01:06.178210 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzhx" event={"ID":"05f88770-9339-4cd6-8aca-417259a85096","Type":"ContainerStarted","Data":"e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7"} Dec 09 18:01:06 crc kubenswrapper[4954]: I1209 18:01:06.208932 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jbzhx" podStartSLOduration=3.55567591 podStartE2EDuration="6.208906859s" podCreationTimestamp="2025-12-09 18:01:00 +0000 UTC" firstStartedPulling="2025-12-09 18:01:02.124788087 +0000 UTC m=+3858.512961907" lastFinishedPulling="2025-12-09 18:01:04.778019036 +0000 UTC m=+3861.166192856" observedRunningTime="2025-12-09 18:01:06.20125846 +0000 UTC m=+3862.589432300" watchObservedRunningTime="2025-12-09 18:01:06.208906859 +0000 UTC m=+3862.597080679" Dec 09 18:01:08 crc kubenswrapper[4954]: E1209 18:01:08.127574 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:01:11 crc kubenswrapper[4954]: I1209 18:01:11.016860 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:11 crc kubenswrapper[4954]: I1209 18:01:11.016993 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:11 crc kubenswrapper[4954]: I1209 18:01:11.101377 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:11 crc kubenswrapper[4954]: I1209 18:01:11.121198 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:01:11 crc kubenswrapper[4954]: E1209 18:01:11.121787 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:01:11 crc kubenswrapper[4954]: I1209 18:01:11.282812 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:11 crc kubenswrapper[4954]: I1209 18:01:11.354429 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jbzhx"] Dec 09 18:01:13 crc kubenswrapper[4954]: I1209 18:01:13.255060 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jbzhx" podUID="05f88770-9339-4cd6-8aca-417259a85096" containerName="registry-server" containerID="cri-o://e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7" gracePeriod=2 Dec 09 18:01:13 crc kubenswrapper[4954]: I1209 18:01:13.764832 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:13 crc kubenswrapper[4954]: I1209 18:01:13.903403 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-catalog-content\") pod \"05f88770-9339-4cd6-8aca-417259a85096\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " Dec 09 18:01:13 crc kubenswrapper[4954]: I1209 18:01:13.903873 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fnrp\" (UniqueName: \"kubernetes.io/projected/05f88770-9339-4cd6-8aca-417259a85096-kube-api-access-5fnrp\") pod \"05f88770-9339-4cd6-8aca-417259a85096\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " Dec 09 18:01:13 crc kubenswrapper[4954]: I1209 18:01:13.904109 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-utilities\") pod \"05f88770-9339-4cd6-8aca-417259a85096\" (UID: \"05f88770-9339-4cd6-8aca-417259a85096\") " Dec 09 18:01:13 crc kubenswrapper[4954]: I1209 18:01:13.906190 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-utilities" (OuterVolumeSpecName: "utilities") pod "05f88770-9339-4cd6-8aca-417259a85096" (UID: "05f88770-9339-4cd6-8aca-417259a85096"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:01:13 crc kubenswrapper[4954]: I1209 18:01:13.906724 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:01:13 crc kubenswrapper[4954]: I1209 18:01:13.915937 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05f88770-9339-4cd6-8aca-417259a85096-kube-api-access-5fnrp" (OuterVolumeSpecName: "kube-api-access-5fnrp") pod "05f88770-9339-4cd6-8aca-417259a85096" (UID: "05f88770-9339-4cd6-8aca-417259a85096"). InnerVolumeSpecName "kube-api-access-5fnrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:01:13 crc kubenswrapper[4954]: I1209 18:01:13.975293 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05f88770-9339-4cd6-8aca-417259a85096" (UID: "05f88770-9339-4cd6-8aca-417259a85096"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.008718 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fnrp\" (UniqueName: \"kubernetes.io/projected/05f88770-9339-4cd6-8aca-417259a85096-kube-api-access-5fnrp\") on node \"crc\" DevicePath \"\"" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.009563 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05f88770-9339-4cd6-8aca-417259a85096-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.270219 4954 generic.go:334] "Generic (PLEG): container finished" podID="05f88770-9339-4cd6-8aca-417259a85096" containerID="e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7" exitCode=0 Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.270269 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzhx" event={"ID":"05f88770-9339-4cd6-8aca-417259a85096","Type":"ContainerDied","Data":"e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7"} Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.270317 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzhx" event={"ID":"05f88770-9339-4cd6-8aca-417259a85096","Type":"ContainerDied","Data":"43322653bc55d981676219d5929a9bd572069e90a1290ddaccd542a05d88f689"} Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.270341 4954 scope.go:117] "RemoveContainer" containerID="e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.270415 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbzhx" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.297730 4954 scope.go:117] "RemoveContainer" containerID="26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.302227 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jbzhx"] Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.318807 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jbzhx"] Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.322139 4954 scope.go:117] "RemoveContainer" containerID="6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.380757 4954 scope.go:117] "RemoveContainer" containerID="e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7" Dec 09 18:01:14 crc kubenswrapper[4954]: E1209 18:01:14.381216 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7\": container with ID starting with e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7 not found: ID does not exist" containerID="e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.381256 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7"} err="failed to get container status \"e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7\": rpc error: code = NotFound desc = could not find container \"e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7\": container with ID starting with e81f7fdb47b52bac8a255aeb41dbf2d270bebbd9d8b5b2f11a4e038d686145c7 not found: ID does not exist" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.381280 4954 scope.go:117] "RemoveContainer" containerID="26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e" Dec 09 18:01:14 crc kubenswrapper[4954]: E1209 18:01:14.381859 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e\": container with ID starting with 26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e not found: ID does not exist" containerID="26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.381906 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e"} err="failed to get container status \"26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e\": rpc error: code = NotFound desc = could not find container \"26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e\": container with ID starting with 26afbea25a240b7a20dcac8a29831d73d1a0219c5272c1b9eaabe986645b4e3e not found: ID does not exist" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.381990 4954 scope.go:117] "RemoveContainer" containerID="6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a" Dec 09 18:01:14 crc kubenswrapper[4954]: E1209 18:01:14.382332 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a\": container with ID starting with 6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a not found: ID does not exist" containerID="6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a" Dec 09 18:01:14 crc kubenswrapper[4954]: I1209 18:01:14.382385 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a"} err="failed to get container status \"6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a\": rpc error: code = NotFound desc = could not find container \"6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a\": container with ID starting with 6b46c0cdcda7bac2847f2a26e4d2949f809524424cfe36d69928a4070c66874a not found: ID does not exist" Dec 09 18:01:16 crc kubenswrapper[4954]: E1209 18:01:16.122472 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:01:16 crc kubenswrapper[4954]: I1209 18:01:16.137404 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05f88770-9339-4cd6-8aca-417259a85096" path="/var/lib/kubelet/pods/05f88770-9339-4cd6-8aca-417259a85096/volumes" Dec 09 18:01:22 crc kubenswrapper[4954]: E1209 18:01:22.122654 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:01:26 crc kubenswrapper[4954]: I1209 18:01:26.120847 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:01:26 crc kubenswrapper[4954]: E1209 18:01:26.123573 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:01:27 crc kubenswrapper[4954]: E1209 18:01:27.128823 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:01:34 crc kubenswrapper[4954]: E1209 18:01:34.129507 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:01:39 crc kubenswrapper[4954]: E1209 18:01:39.123917 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:01:40 crc kubenswrapper[4954]: I1209 18:01:40.120925 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:01:40 crc kubenswrapper[4954]: E1209 18:01:40.121824 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:01:46 crc kubenswrapper[4954]: E1209 18:01:46.122491 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:01:52 crc kubenswrapper[4954]: E1209 18:01:52.122410 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:01:55 crc kubenswrapper[4954]: I1209 18:01:55.120437 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:01:55 crc kubenswrapper[4954]: E1209 18:01:55.121053 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:01:57 crc kubenswrapper[4954]: E1209 18:01:57.123981 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.053781 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq"] Dec 09 18:02:06 crc kubenswrapper[4954]: E1209 18:02:06.055015 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f88770-9339-4cd6-8aca-417259a85096" containerName="extract-utilities" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.055035 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f88770-9339-4cd6-8aca-417259a85096" containerName="extract-utilities" Dec 09 18:02:06 crc kubenswrapper[4954]: E1209 18:02:06.055069 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d" containerName="keystone-cron" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.055078 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d" containerName="keystone-cron" Dec 09 18:02:06 crc kubenswrapper[4954]: E1209 18:02:06.055095 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f88770-9339-4cd6-8aca-417259a85096" containerName="extract-content" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.055103 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f88770-9339-4cd6-8aca-417259a85096" containerName="extract-content" Dec 09 18:02:06 crc kubenswrapper[4954]: E1209 18:02:06.055127 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f88770-9339-4cd6-8aca-417259a85096" containerName="registry-server" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.055135 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f88770-9339-4cd6-8aca-417259a85096" containerName="registry-server" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.055417 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="05f88770-9339-4cd6-8aca-417259a85096" containerName="registry-server" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.055450 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d" containerName="keystone-cron" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.065840 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.068871 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq"] Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.068915 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.069203 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.069257 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7697j" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.069475 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.069823 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.200823 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65r5s\" (UniqueName: \"kubernetes.io/projected/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-kube-api-access-65r5s\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.200873 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.200933 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.200956 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.201106 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.201411 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.201476 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.303693 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65r5s\" (UniqueName: \"kubernetes.io/projected/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-kube-api-access-65r5s\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.303752 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.303843 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.304520 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.304789 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.305292 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.305346 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.310552 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.310573 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.310771 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.311321 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.311424 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.311473 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.327296 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65r5s\" (UniqueName: \"kubernetes.io/projected/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-kube-api-access-65r5s\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.387369 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:02:06 crc kubenswrapper[4954]: W1209 18:02:06.933848 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d2dcb4c_f20a_467a_9e39_4577f464c3f1.slice/crio-36a8d6e9a0f3903a7d1bee81081e36c3929210ab0c598ca1944205f8fe289cb3 WatchSource:0}: Error finding container 36a8d6e9a0f3903a7d1bee81081e36c3929210ab0c598ca1944205f8fe289cb3: Status 404 returned error can't find the container with id 36a8d6e9a0f3903a7d1bee81081e36c3929210ab0c598ca1944205f8fe289cb3 Dec 09 18:02:06 crc kubenswrapper[4954]: I1209 18:02:06.938809 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq"] Dec 09 18:02:07 crc kubenswrapper[4954]: I1209 18:02:07.120714 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:02:07 crc kubenswrapper[4954]: E1209 18:02:07.121290 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:02:07 crc kubenswrapper[4954]: E1209 18:02:07.123419 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:02:07 crc kubenswrapper[4954]: I1209 18:02:07.803479 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" event={"ID":"9d2dcb4c-f20a-467a-9e39-4577f464c3f1","Type":"ContainerStarted","Data":"35b00feda9774a4fe577a3b6b822d4171244ea26d5834771be9146ea2d5e5e83"} Dec 09 18:02:07 crc kubenswrapper[4954]: I1209 18:02:07.803856 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" event={"ID":"9d2dcb4c-f20a-467a-9e39-4577f464c3f1","Type":"ContainerStarted","Data":"36a8d6e9a0f3903a7d1bee81081e36c3929210ab0c598ca1944205f8fe289cb3"} Dec 09 18:02:07 crc kubenswrapper[4954]: I1209 18:02:07.832424 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" podStartSLOduration=1.321820469 podStartE2EDuration="1.832401182s" podCreationTimestamp="2025-12-09 18:02:06 +0000 UTC" firstStartedPulling="2025-12-09 18:02:06.936433724 +0000 UTC m=+3923.324607544" lastFinishedPulling="2025-12-09 18:02:07.447014427 +0000 UTC m=+3923.835188257" observedRunningTime="2025-12-09 18:02:07.821010566 +0000 UTC m=+3924.209184396" watchObservedRunningTime="2025-12-09 18:02:07.832401182 +0000 UTC m=+3924.220575002" Dec 09 18:02:09 crc kubenswrapper[4954]: E1209 18:02:09.122584 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:02:20 crc kubenswrapper[4954]: I1209 18:02:20.120887 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:02:20 crc kubenswrapper[4954]: E1209 18:02:20.121754 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:02:21 crc kubenswrapper[4954]: E1209 18:02:21.122470 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:02:21 crc kubenswrapper[4954]: E1209 18:02:21.122507 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:02:32 crc kubenswrapper[4954]: E1209 18:02:32.127572 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:02:34 crc kubenswrapper[4954]: I1209 18:02:34.128427 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:02:34 crc kubenswrapper[4954]: E1209 18:02:34.129270 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:02:36 crc kubenswrapper[4954]: E1209 18:02:36.122194 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:02:43 crc kubenswrapper[4954]: E1209 18:02:43.122950 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:02:46 crc kubenswrapper[4954]: I1209 18:02:46.121283 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:02:46 crc kubenswrapper[4954]: E1209 18:02:46.123422 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:02:48 crc kubenswrapper[4954]: I1209 18:02:48.123094 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 18:02:48 crc kubenswrapper[4954]: E1209 18:02:48.254667 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:02:48 crc kubenswrapper[4954]: E1209 18:02:48.254771 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:02:48 crc kubenswrapper[4954]: E1209 18:02:48.255066 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:02:48 crc kubenswrapper[4954]: E1209 18:02:48.256234 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:02:56 crc kubenswrapper[4954]: E1209 18:02:56.123169 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:02:59 crc kubenswrapper[4954]: I1209 18:02:59.120415 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:02:59 crc kubenswrapper[4954]: E1209 18:02:59.121232 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:03:01 crc kubenswrapper[4954]: E1209 18:03:01.121915 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:03:06 crc kubenswrapper[4954]: I1209 18:03:06.361858 4954 generic.go:334] "Generic (PLEG): container finished" podID="9d2dcb4c-f20a-467a-9e39-4577f464c3f1" containerID="35b00feda9774a4fe577a3b6b822d4171244ea26d5834771be9146ea2d5e5e83" exitCode=2 Dec 09 18:03:06 crc kubenswrapper[4954]: I1209 18:03:06.361972 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" event={"ID":"9d2dcb4c-f20a-467a-9e39-4577f464c3f1","Type":"ContainerDied","Data":"35b00feda9774a4fe577a3b6b822d4171244ea26d5834771be9146ea2d5e5e83"} Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.848954 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.949255 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ssh-key\") pod \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.949316 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65r5s\" (UniqueName: \"kubernetes.io/projected/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-kube-api-access-65r5s\") pod \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.949352 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-2\") pod \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.949395 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-inventory\") pod \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.949526 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-telemetry-combined-ca-bundle\") pod \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.949668 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-1\") pod \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.949737 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-0\") pod \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\" (UID: \"9d2dcb4c-f20a-467a-9e39-4577f464c3f1\") " Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.954939 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-kube-api-access-65r5s" (OuterVolumeSpecName: "kube-api-access-65r5s") pod "9d2dcb4c-f20a-467a-9e39-4577f464c3f1" (UID: "9d2dcb4c-f20a-467a-9e39-4577f464c3f1"). InnerVolumeSpecName "kube-api-access-65r5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.962014 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "9d2dcb4c-f20a-467a-9e39-4577f464c3f1" (UID: "9d2dcb4c-f20a-467a-9e39-4577f464c3f1"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.982214 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "9d2dcb4c-f20a-467a-9e39-4577f464c3f1" (UID: "9d2dcb4c-f20a-467a-9e39-4577f464c3f1"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.985205 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-inventory" (OuterVolumeSpecName: "inventory") pod "9d2dcb4c-f20a-467a-9e39-4577f464c3f1" (UID: "9d2dcb4c-f20a-467a-9e39-4577f464c3f1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.987120 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9d2dcb4c-f20a-467a-9e39-4577f464c3f1" (UID: "9d2dcb4c-f20a-467a-9e39-4577f464c3f1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:03:07 crc kubenswrapper[4954]: I1209 18:03:07.987415 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "9d2dcb4c-f20a-467a-9e39-4577f464c3f1" (UID: "9d2dcb4c-f20a-467a-9e39-4577f464c3f1"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.002760 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "9d2dcb4c-f20a-467a-9e39-4577f464c3f1" (UID: "9d2dcb4c-f20a-467a-9e39-4577f464c3f1"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.052223 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65r5s\" (UniqueName: \"kubernetes.io/projected/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-kube-api-access-65r5s\") on node \"crc\" DevicePath \"\"" Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.052257 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.052270 4954 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.052280 4954 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.052292 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.052301 4954 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.052310 4954 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d2dcb4c-f20a-467a-9e39-4577f464c3f1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 18:03:08 crc kubenswrapper[4954]: E1209 18:03:08.233749 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:03:08 crc kubenswrapper[4954]: E1209 18:03:08.234020 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:03:08 crc kubenswrapper[4954]: E1209 18:03:08.234244 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:03:08 crc kubenswrapper[4954]: E1209 18:03:08.235537 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.399110 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.398937 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq" event={"ID":"9d2dcb4c-f20a-467a-9e39-4577f464c3f1","Type":"ContainerDied","Data":"36a8d6e9a0f3903a7d1bee81081e36c3929210ab0c598ca1944205f8fe289cb3"} Dec 09 18:03:08 crc kubenswrapper[4954]: I1209 18:03:08.399572 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36a8d6e9a0f3903a7d1bee81081e36c3929210ab0c598ca1944205f8fe289cb3" Dec 09 18:03:12 crc kubenswrapper[4954]: I1209 18:03:12.120664 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:03:12 crc kubenswrapper[4954]: E1209 18:03:12.121441 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:03:12 crc kubenswrapper[4954]: E1209 18:03:12.124151 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:03:22 crc kubenswrapper[4954]: E1209 18:03:22.124480 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:03:23 crc kubenswrapper[4954]: I1209 18:03:23.121807 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:03:23 crc kubenswrapper[4954]: E1209 18:03:23.122112 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:03:24 crc kubenswrapper[4954]: E1209 18:03:24.129576 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:03:36 crc kubenswrapper[4954]: E1209 18:03:36.123642 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:03:37 crc kubenswrapper[4954]: I1209 18:03:37.122009 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:03:37 crc kubenswrapper[4954]: E1209 18:03:37.122658 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:03:38 crc kubenswrapper[4954]: E1209 18:03:38.124413 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:03:48 crc kubenswrapper[4954]: E1209 18:03:48.125692 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:03:50 crc kubenswrapper[4954]: E1209 18:03:50.123539 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:03:52 crc kubenswrapper[4954]: I1209 18:03:52.119742 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:03:52 crc kubenswrapper[4954]: E1209 18:03:52.120355 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:03:59 crc kubenswrapper[4954]: E1209 18:03:59.121992 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:04:05 crc kubenswrapper[4954]: E1209 18:04:05.123208 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:04:07 crc kubenswrapper[4954]: I1209 18:04:07.121212 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:04:07 crc kubenswrapper[4954]: E1209 18:04:07.122060 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:04:10 crc kubenswrapper[4954]: E1209 18:04:10.124071 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.120554 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:04:19 crc kubenswrapper[4954]: E1209 18:04:19.121476 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.379937 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jvp7v"] Dec 09 18:04:19 crc kubenswrapper[4954]: E1209 18:04:19.380464 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d2dcb4c-f20a-467a-9e39-4577f464c3f1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.380486 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d2dcb4c-f20a-467a-9e39-4577f464c3f1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.380953 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d2dcb4c-f20a-467a-9e39-4577f464c3f1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.383093 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.393561 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvp7v"] Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.404642 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-utilities\") pod \"redhat-marketplace-jvp7v\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.404685 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-catalog-content\") pod \"redhat-marketplace-jvp7v\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.404761 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7mjx\" (UniqueName: \"kubernetes.io/projected/78e569c6-6c33-4494-a609-6217e4095821-kube-api-access-r7mjx\") pod \"redhat-marketplace-jvp7v\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.507008 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-utilities\") pod \"redhat-marketplace-jvp7v\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.507058 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-catalog-content\") pod \"redhat-marketplace-jvp7v\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.507127 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7mjx\" (UniqueName: \"kubernetes.io/projected/78e569c6-6c33-4494-a609-6217e4095821-kube-api-access-r7mjx\") pod \"redhat-marketplace-jvp7v\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.507675 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-utilities\") pod \"redhat-marketplace-jvp7v\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.507792 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-catalog-content\") pod \"redhat-marketplace-jvp7v\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.526492 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7mjx\" (UniqueName: \"kubernetes.io/projected/78e569c6-6c33-4494-a609-6217e4095821-kube-api-access-r7mjx\") pod \"redhat-marketplace-jvp7v\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:19 crc kubenswrapper[4954]: I1209 18:04:19.709680 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:20 crc kubenswrapper[4954]: E1209 18:04:20.124264 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:04:20 crc kubenswrapper[4954]: I1209 18:04:20.198158 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvp7v"] Dec 09 18:04:21 crc kubenswrapper[4954]: I1209 18:04:21.084422 4954 generic.go:334] "Generic (PLEG): container finished" podID="78e569c6-6c33-4494-a609-6217e4095821" containerID="5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0" exitCode=0 Dec 09 18:04:21 crc kubenswrapper[4954]: I1209 18:04:21.084807 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvp7v" event={"ID":"78e569c6-6c33-4494-a609-6217e4095821","Type":"ContainerDied","Data":"5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0"} Dec 09 18:04:21 crc kubenswrapper[4954]: I1209 18:04:21.084830 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvp7v" event={"ID":"78e569c6-6c33-4494-a609-6217e4095821","Type":"ContainerStarted","Data":"fa176d7a475c5e6065051a19c51f436deb63545a8deb08e5d6de89a6904dcd3e"} Dec 09 18:04:23 crc kubenswrapper[4954]: I1209 18:04:23.105052 4954 generic.go:334] "Generic (PLEG): container finished" podID="78e569c6-6c33-4494-a609-6217e4095821" containerID="95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77" exitCode=0 Dec 09 18:04:23 crc kubenswrapper[4954]: I1209 18:04:23.105156 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvp7v" event={"ID":"78e569c6-6c33-4494-a609-6217e4095821","Type":"ContainerDied","Data":"95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77"} Dec 09 18:04:24 crc kubenswrapper[4954]: I1209 18:04:24.117389 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvp7v" event={"ID":"78e569c6-6c33-4494-a609-6217e4095821","Type":"ContainerStarted","Data":"8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997"} Dec 09 18:04:24 crc kubenswrapper[4954]: E1209 18:04:24.122353 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:04:24 crc kubenswrapper[4954]: I1209 18:04:24.159398 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jvp7v" podStartSLOduration=2.625298031 podStartE2EDuration="5.159370169s" podCreationTimestamp="2025-12-09 18:04:19 +0000 UTC" firstStartedPulling="2025-12-09 18:04:21.086967678 +0000 UTC m=+4057.475141498" lastFinishedPulling="2025-12-09 18:04:23.621039816 +0000 UTC m=+4060.009213636" observedRunningTime="2025-12-09 18:04:24.149218472 +0000 UTC m=+4060.537392302" watchObservedRunningTime="2025-12-09 18:04:24.159370169 +0000 UTC m=+4060.547543989" Dec 09 18:04:29 crc kubenswrapper[4954]: I1209 18:04:29.715582 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:29 crc kubenswrapper[4954]: I1209 18:04:29.716215 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:29 crc kubenswrapper[4954]: I1209 18:04:29.765918 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:30 crc kubenswrapper[4954]: I1209 18:04:30.212299 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:30 crc kubenswrapper[4954]: I1209 18:04:30.256193 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvp7v"] Dec 09 18:04:31 crc kubenswrapper[4954]: I1209 18:04:31.120219 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:04:31 crc kubenswrapper[4954]: E1209 18:04:31.120561 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:04:32 crc kubenswrapper[4954]: E1209 18:04:32.124231 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:04:32 crc kubenswrapper[4954]: I1209 18:04:32.189259 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jvp7v" podUID="78e569c6-6c33-4494-a609-6217e4095821" containerName="registry-server" containerID="cri-o://8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997" gracePeriod=2 Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.151686 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.201930 4954 generic.go:334] "Generic (PLEG): container finished" podID="78e569c6-6c33-4494-a609-6217e4095821" containerID="8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997" exitCode=0 Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.201967 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvp7v" event={"ID":"78e569c6-6c33-4494-a609-6217e4095821","Type":"ContainerDied","Data":"8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997"} Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.201992 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvp7v" event={"ID":"78e569c6-6c33-4494-a609-6217e4095821","Type":"ContainerDied","Data":"fa176d7a475c5e6065051a19c51f436deb63545a8deb08e5d6de89a6904dcd3e"} Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.202011 4954 scope.go:117] "RemoveContainer" containerID="8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.201994 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvp7v" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.208818 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-utilities\") pod \"78e569c6-6c33-4494-a609-6217e4095821\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.208984 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-catalog-content\") pod \"78e569c6-6c33-4494-a609-6217e4095821\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.209123 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7mjx\" (UniqueName: \"kubernetes.io/projected/78e569c6-6c33-4494-a609-6217e4095821-kube-api-access-r7mjx\") pod \"78e569c6-6c33-4494-a609-6217e4095821\" (UID: \"78e569c6-6c33-4494-a609-6217e4095821\") " Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.209788 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-utilities" (OuterVolumeSpecName: "utilities") pod "78e569c6-6c33-4494-a609-6217e4095821" (UID: "78e569c6-6c33-4494-a609-6217e4095821"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.214663 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78e569c6-6c33-4494-a609-6217e4095821-kube-api-access-r7mjx" (OuterVolumeSpecName: "kube-api-access-r7mjx") pod "78e569c6-6c33-4494-a609-6217e4095821" (UID: "78e569c6-6c33-4494-a609-6217e4095821"). InnerVolumeSpecName "kube-api-access-r7mjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.227233 4954 scope.go:117] "RemoveContainer" containerID="95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.228904 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78e569c6-6c33-4494-a609-6217e4095821" (UID: "78e569c6-6c33-4494-a609-6217e4095821"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.279682 4954 scope.go:117] "RemoveContainer" containerID="5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.319695 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7mjx\" (UniqueName: \"kubernetes.io/projected/78e569c6-6c33-4494-a609-6217e4095821-kube-api-access-r7mjx\") on node \"crc\" DevicePath \"\"" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.319748 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.319762 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78e569c6-6c33-4494-a609-6217e4095821-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.343958 4954 scope.go:117] "RemoveContainer" containerID="8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997" Dec 09 18:04:33 crc kubenswrapper[4954]: E1209 18:04:33.344354 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997\": container with ID starting with 8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997 not found: ID does not exist" containerID="8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.344384 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997"} err="failed to get container status \"8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997\": rpc error: code = NotFound desc = could not find container \"8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997\": container with ID starting with 8e4a773a2afd4cef0dbccbb3aae2f8b00c88128560379cefb98c931061c86997 not found: ID does not exist" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.344406 4954 scope.go:117] "RemoveContainer" containerID="95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77" Dec 09 18:04:33 crc kubenswrapper[4954]: E1209 18:04:33.345131 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77\": container with ID starting with 95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77 not found: ID does not exist" containerID="95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.345151 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77"} err="failed to get container status \"95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77\": rpc error: code = NotFound desc = could not find container \"95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77\": container with ID starting with 95c2b12edd1e67f40cfe5c4590e231e2813f1eb81af73212f51bb8b30c02bb77 not found: ID does not exist" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.345162 4954 scope.go:117] "RemoveContainer" containerID="5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0" Dec 09 18:04:33 crc kubenswrapper[4954]: E1209 18:04:33.345410 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0\": container with ID starting with 5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0 not found: ID does not exist" containerID="5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.345430 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0"} err="failed to get container status \"5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0\": rpc error: code = NotFound desc = could not find container \"5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0\": container with ID starting with 5a0b62c1809465f36041a7455cdaa13a83312b2f7c08a42c4b7b3f85219f44c0 not found: ID does not exist" Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.540168 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvp7v"] Dec 09 18:04:33 crc kubenswrapper[4954]: I1209 18:04:33.551076 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvp7v"] Dec 09 18:04:34 crc kubenswrapper[4954]: I1209 18:04:34.135924 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78e569c6-6c33-4494-a609-6217e4095821" path="/var/lib/kubelet/pods/78e569c6-6c33-4494-a609-6217e4095821/volumes" Dec 09 18:04:38 crc kubenswrapper[4954]: E1209 18:04:38.122167 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:04:43 crc kubenswrapper[4954]: E1209 18:04:43.123682 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:04:44 crc kubenswrapper[4954]: I1209 18:04:44.128551 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:04:44 crc kubenswrapper[4954]: E1209 18:04:44.129496 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:04:51 crc kubenswrapper[4954]: E1209 18:04:51.124917 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:04:54 crc kubenswrapper[4954]: E1209 18:04:54.132475 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:04:58 crc kubenswrapper[4954]: I1209 18:04:58.121402 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:04:58 crc kubenswrapper[4954]: E1209 18:04:58.121944 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:05:02 crc kubenswrapper[4954]: E1209 18:05:02.129790 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:05:09 crc kubenswrapper[4954]: E1209 18:05:09.122799 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:05:11 crc kubenswrapper[4954]: I1209 18:05:11.120944 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:05:11 crc kubenswrapper[4954]: E1209 18:05:11.121555 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:05:15 crc kubenswrapper[4954]: E1209 18:05:15.123457 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:05:23 crc kubenswrapper[4954]: E1209 18:05:23.123063 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:05:25 crc kubenswrapper[4954]: I1209 18:05:25.901635 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jmg69"] Dec 09 18:05:25 crc kubenswrapper[4954]: E1209 18:05:25.902675 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78e569c6-6c33-4494-a609-6217e4095821" containerName="registry-server" Dec 09 18:05:25 crc kubenswrapper[4954]: I1209 18:05:25.902693 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="78e569c6-6c33-4494-a609-6217e4095821" containerName="registry-server" Dec 09 18:05:25 crc kubenswrapper[4954]: E1209 18:05:25.902712 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78e569c6-6c33-4494-a609-6217e4095821" containerName="extract-content" Dec 09 18:05:25 crc kubenswrapper[4954]: I1209 18:05:25.902720 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="78e569c6-6c33-4494-a609-6217e4095821" containerName="extract-content" Dec 09 18:05:25 crc kubenswrapper[4954]: E1209 18:05:25.902749 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78e569c6-6c33-4494-a609-6217e4095821" containerName="extract-utilities" Dec 09 18:05:25 crc kubenswrapper[4954]: I1209 18:05:25.902758 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="78e569c6-6c33-4494-a609-6217e4095821" containerName="extract-utilities" Dec 09 18:05:25 crc kubenswrapper[4954]: I1209 18:05:25.903016 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="78e569c6-6c33-4494-a609-6217e4095821" containerName="registry-server" Dec 09 18:05:25 crc kubenswrapper[4954]: I1209 18:05:25.904989 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:25 crc kubenswrapper[4954]: I1209 18:05:25.917404 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jmg69"] Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.049570 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9ln5\" (UniqueName: \"kubernetes.io/projected/a17b20b0-c81f-42c5-96fb-19554d40f90b-kube-api-access-r9ln5\") pod \"redhat-operators-jmg69\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.049710 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-utilities\") pod \"redhat-operators-jmg69\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.049757 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-catalog-content\") pod \"redhat-operators-jmg69\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.120460 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:05:26 crc kubenswrapper[4954]: E1209 18:05:26.133074 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:05:26 crc kubenswrapper[4954]: E1209 18:05:26.136502 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.156469 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-utilities\") pod \"redhat-operators-jmg69\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.156531 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-catalog-content\") pod \"redhat-operators-jmg69\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.156841 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9ln5\" (UniqueName: \"kubernetes.io/projected/a17b20b0-c81f-42c5-96fb-19554d40f90b-kube-api-access-r9ln5\") pod \"redhat-operators-jmg69\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.157066 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-utilities\") pod \"redhat-operators-jmg69\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.157277 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-catalog-content\") pod \"redhat-operators-jmg69\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.200153 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9ln5\" (UniqueName: \"kubernetes.io/projected/a17b20b0-c81f-42c5-96fb-19554d40f90b-kube-api-access-r9ln5\") pod \"redhat-operators-jmg69\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.233314 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:26 crc kubenswrapper[4954]: I1209 18:05:26.777069 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jmg69"] Dec 09 18:05:27 crc kubenswrapper[4954]: I1209 18:05:27.723813 4954 generic.go:334] "Generic (PLEG): container finished" podID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerID="b11242f90e029d5732ffec39bd36e13d2fffd1148ee1721883f8dce9e0c54e16" exitCode=0 Dec 09 18:05:27 crc kubenswrapper[4954]: I1209 18:05:27.723915 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmg69" event={"ID":"a17b20b0-c81f-42c5-96fb-19554d40f90b","Type":"ContainerDied","Data":"b11242f90e029d5732ffec39bd36e13d2fffd1148ee1721883f8dce9e0c54e16"} Dec 09 18:05:27 crc kubenswrapper[4954]: I1209 18:05:27.724148 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmg69" event={"ID":"a17b20b0-c81f-42c5-96fb-19554d40f90b","Type":"ContainerStarted","Data":"d891458a6ecf3d6c3e8ab08275a27729e91c8667601180464bb95d3b43b5717a"} Dec 09 18:05:30 crc kubenswrapper[4954]: I1209 18:05:30.764440 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmg69" event={"ID":"a17b20b0-c81f-42c5-96fb-19554d40f90b","Type":"ContainerStarted","Data":"6ecfc7fb7cd54dbc6a9f41cd7c362c07a27cc115b06bb79adbb0878ab4640ca6"} Dec 09 18:05:33 crc kubenswrapper[4954]: I1209 18:05:33.795464 4954 generic.go:334] "Generic (PLEG): container finished" podID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerID="6ecfc7fb7cd54dbc6a9f41cd7c362c07a27cc115b06bb79adbb0878ab4640ca6" exitCode=0 Dec 09 18:05:33 crc kubenswrapper[4954]: I1209 18:05:33.795544 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmg69" event={"ID":"a17b20b0-c81f-42c5-96fb-19554d40f90b","Type":"ContainerDied","Data":"6ecfc7fb7cd54dbc6a9f41cd7c362c07a27cc115b06bb79adbb0878ab4640ca6"} Dec 09 18:05:34 crc kubenswrapper[4954]: E1209 18:05:34.131986 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:05:34 crc kubenswrapper[4954]: I1209 18:05:34.809934 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmg69" event={"ID":"a17b20b0-c81f-42c5-96fb-19554d40f90b","Type":"ContainerStarted","Data":"34f1ebcaf5ec702324068294d3dd0e818a929bdbcb373aa5c47a0981a14e20a4"} Dec 09 18:05:34 crc kubenswrapper[4954]: I1209 18:05:34.836797 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jmg69" podStartSLOduration=3.117560072 podStartE2EDuration="9.836777824s" podCreationTimestamp="2025-12-09 18:05:25 +0000 UTC" firstStartedPulling="2025-12-09 18:05:27.725818099 +0000 UTC m=+4124.113991919" lastFinishedPulling="2025-12-09 18:05:34.445035851 +0000 UTC m=+4130.833209671" observedRunningTime="2025-12-09 18:05:34.826529414 +0000 UTC m=+4131.214703264" watchObservedRunningTime="2025-12-09 18:05:34.836777824 +0000 UTC m=+4131.224951644" Dec 09 18:05:36 crc kubenswrapper[4954]: I1209 18:05:36.234254 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:36 crc kubenswrapper[4954]: I1209 18:05:36.234633 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:37 crc kubenswrapper[4954]: I1209 18:05:37.287923 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jmg69" podUID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerName="registry-server" probeResult="failure" output=< Dec 09 18:05:37 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 18:05:37 crc kubenswrapper[4954]: > Dec 09 18:05:40 crc kubenswrapper[4954]: E1209 18:05:40.123904 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:05:40 crc kubenswrapper[4954]: I1209 18:05:40.133206 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:05:40 crc kubenswrapper[4954]: E1209 18:05:40.133678 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:05:46 crc kubenswrapper[4954]: I1209 18:05:46.289275 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:46 crc kubenswrapper[4954]: I1209 18:05:46.340402 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:47 crc kubenswrapper[4954]: E1209 18:05:47.123113 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:05:50 crc kubenswrapper[4954]: I1209 18:05:50.407951 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jmg69"] Dec 09 18:05:50 crc kubenswrapper[4954]: I1209 18:05:50.408766 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jmg69" podUID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerName="registry-server" containerID="cri-o://34f1ebcaf5ec702324068294d3dd0e818a929bdbcb373aa5c47a0981a14e20a4" gracePeriod=2 Dec 09 18:05:50 crc kubenswrapper[4954]: I1209 18:05:50.957584 4954 generic.go:334] "Generic (PLEG): container finished" podID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerID="34f1ebcaf5ec702324068294d3dd0e818a929bdbcb373aa5c47a0981a14e20a4" exitCode=0 Dec 09 18:05:50 crc kubenswrapper[4954]: I1209 18:05:50.957807 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmg69" event={"ID":"a17b20b0-c81f-42c5-96fb-19554d40f90b","Type":"ContainerDied","Data":"34f1ebcaf5ec702324068294d3dd0e818a929bdbcb373aa5c47a0981a14e20a4"} Dec 09 18:05:50 crc kubenswrapper[4954]: I1209 18:05:50.957897 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jmg69" event={"ID":"a17b20b0-c81f-42c5-96fb-19554d40f90b","Type":"ContainerDied","Data":"d891458a6ecf3d6c3e8ab08275a27729e91c8667601180464bb95d3b43b5717a"} Dec 09 18:05:50 crc kubenswrapper[4954]: I1209 18:05:50.957912 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d891458a6ecf3d6c3e8ab08275a27729e91c8667601180464bb95d3b43b5717a" Dec 09 18:05:50 crc kubenswrapper[4954]: I1209 18:05:50.972033 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:51 crc kubenswrapper[4954]: I1209 18:05:51.123467 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-catalog-content\") pod \"a17b20b0-c81f-42c5-96fb-19554d40f90b\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " Dec 09 18:05:51 crc kubenswrapper[4954]: I1209 18:05:51.123544 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-utilities\") pod \"a17b20b0-c81f-42c5-96fb-19554d40f90b\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " Dec 09 18:05:51 crc kubenswrapper[4954]: I1209 18:05:51.123719 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9ln5\" (UniqueName: \"kubernetes.io/projected/a17b20b0-c81f-42c5-96fb-19554d40f90b-kube-api-access-r9ln5\") pod \"a17b20b0-c81f-42c5-96fb-19554d40f90b\" (UID: \"a17b20b0-c81f-42c5-96fb-19554d40f90b\") " Dec 09 18:05:51 crc kubenswrapper[4954]: I1209 18:05:51.124632 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-utilities" (OuterVolumeSpecName: "utilities") pod "a17b20b0-c81f-42c5-96fb-19554d40f90b" (UID: "a17b20b0-c81f-42c5-96fb-19554d40f90b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:05:51 crc kubenswrapper[4954]: I1209 18:05:51.131952 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a17b20b0-c81f-42c5-96fb-19554d40f90b-kube-api-access-r9ln5" (OuterVolumeSpecName: "kube-api-access-r9ln5") pod "a17b20b0-c81f-42c5-96fb-19554d40f90b" (UID: "a17b20b0-c81f-42c5-96fb-19554d40f90b"). InnerVolumeSpecName "kube-api-access-r9ln5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:05:51 crc kubenswrapper[4954]: I1209 18:05:51.228152 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:05:51 crc kubenswrapper[4954]: I1209 18:05:51.228277 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9ln5\" (UniqueName: \"kubernetes.io/projected/a17b20b0-c81f-42c5-96fb-19554d40f90b-kube-api-access-r9ln5\") on node \"crc\" DevicePath \"\"" Dec 09 18:05:51 crc kubenswrapper[4954]: I1209 18:05:51.244578 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a17b20b0-c81f-42c5-96fb-19554d40f90b" (UID: "a17b20b0-c81f-42c5-96fb-19554d40f90b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:05:51 crc kubenswrapper[4954]: I1209 18:05:51.330732 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17b20b0-c81f-42c5-96fb-19554d40f90b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:05:51 crc kubenswrapper[4954]: I1209 18:05:51.967157 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jmg69" Dec 09 18:05:52 crc kubenswrapper[4954]: I1209 18:05:52.010473 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jmg69"] Dec 09 18:05:52 crc kubenswrapper[4954]: I1209 18:05:52.021657 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jmg69"] Dec 09 18:05:52 crc kubenswrapper[4954]: I1209 18:05:52.131428 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a17b20b0-c81f-42c5-96fb-19554d40f90b" path="/var/lib/kubelet/pods/a17b20b0-c81f-42c5-96fb-19554d40f90b/volumes" Dec 09 18:05:54 crc kubenswrapper[4954]: I1209 18:05:54.127505 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:05:55 crc kubenswrapper[4954]: I1209 18:05:55.002619 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"40c79a44d3663bf94a20a1c5aea3afef5e17d945f82090de1b839c4a645080fd"} Dec 09 18:05:55 crc kubenswrapper[4954]: E1209 18:05:55.123479 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:05:59 crc kubenswrapper[4954]: E1209 18:05:59.123777 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:06:10 crc kubenswrapper[4954]: E1209 18:06:10.124320 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:06:10 crc kubenswrapper[4954]: E1209 18:06:10.125622 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:06:24 crc kubenswrapper[4954]: E1209 18:06:24.138781 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:06:25 crc kubenswrapper[4954]: E1209 18:06:25.123110 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:06:36 crc kubenswrapper[4954]: E1209 18:06:36.122480 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:06:40 crc kubenswrapper[4954]: E1209 18:06:40.124028 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:06:51 crc kubenswrapper[4954]: E1209 18:06:51.180845 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:06:51 crc kubenswrapper[4954]: E1209 18:06:51.180870 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:07:05 crc kubenswrapper[4954]: E1209 18:07:05.122085 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:07:06 crc kubenswrapper[4954]: E1209 18:07:06.123793 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:07:18 crc kubenswrapper[4954]: E1209 18:07:18.123245 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:07:19 crc kubenswrapper[4954]: E1209 18:07:19.121798 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:07:29 crc kubenswrapper[4954]: E1209 18:07:29.123426 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:07:34 crc kubenswrapper[4954]: E1209 18:07:34.137248 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:07:44 crc kubenswrapper[4954]: E1209 18:07:44.133701 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:07:47 crc kubenswrapper[4954]: E1209 18:07:47.122396 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:07:59 crc kubenswrapper[4954]: E1209 18:07:59.123412 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:07:59 crc kubenswrapper[4954]: I1209 18:07:59.123694 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 18:07:59 crc kubenswrapper[4954]: E1209 18:07:59.223725 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:07:59 crc kubenswrapper[4954]: E1209 18:07:59.223773 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:07:59 crc kubenswrapper[4954]: E1209 18:07:59.223954 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:07:59 crc kubenswrapper[4954]: E1209 18:07:59.225119 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:08:12 crc kubenswrapper[4954]: E1209 18:08:12.246369 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:08:12 crc kubenswrapper[4954]: E1209 18:08:12.246946 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:08:12 crc kubenswrapper[4954]: E1209 18:08:12.247074 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:08:12 crc kubenswrapper[4954]: E1209 18:08:12.248255 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:08:13 crc kubenswrapper[4954]: E1209 18:08:13.127505 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:08:13 crc kubenswrapper[4954]: I1209 18:08:13.756112 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:08:13 crc kubenswrapper[4954]: I1209 18:08:13.756467 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:08:27 crc kubenswrapper[4954]: E1209 18:08:27.121624 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:08:27 crc kubenswrapper[4954]: E1209 18:08:27.121624 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:08:39 crc kubenswrapper[4954]: E1209 18:08:39.123433 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:08:39 crc kubenswrapper[4954]: E1209 18:08:39.124310 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:08:42 crc kubenswrapper[4954]: I1209 18:08:42.485539 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-547f8767c7-hv854" podUID="a6635c8e-d9a5-4034-ab5c-7fe96dc10c10" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 09 18:08:43 crc kubenswrapper[4954]: I1209 18:08:43.756148 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:08:43 crc kubenswrapper[4954]: I1209 18:08:43.756217 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:08:50 crc kubenswrapper[4954]: E1209 18:08:50.125486 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:08:54 crc kubenswrapper[4954]: E1209 18:08:54.143115 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:09:06 crc kubenswrapper[4954]: E1209 18:09:06.123429 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:09:06 crc kubenswrapper[4954]: E1209 18:09:06.123685 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:09:13 crc kubenswrapper[4954]: I1209 18:09:13.755240 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:09:13 crc kubenswrapper[4954]: I1209 18:09:13.755853 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:09:13 crc kubenswrapper[4954]: I1209 18:09:13.755912 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 18:09:13 crc kubenswrapper[4954]: I1209 18:09:13.756748 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"40c79a44d3663bf94a20a1c5aea3afef5e17d945f82090de1b839c4a645080fd"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 18:09:13 crc kubenswrapper[4954]: I1209 18:09:13.756859 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://40c79a44d3663bf94a20a1c5aea3afef5e17d945f82090de1b839c4a645080fd" gracePeriod=600 Dec 09 18:09:14 crc kubenswrapper[4954]: I1209 18:09:14.120528 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="40c79a44d3663bf94a20a1c5aea3afef5e17d945f82090de1b839c4a645080fd" exitCode=0 Dec 09 18:09:14 crc kubenswrapper[4954]: I1209 18:09:14.133199 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"40c79a44d3663bf94a20a1c5aea3afef5e17d945f82090de1b839c4a645080fd"} Dec 09 18:09:14 crc kubenswrapper[4954]: I1209 18:09:14.133265 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e"} Dec 09 18:09:14 crc kubenswrapper[4954]: I1209 18:09:14.133290 4954 scope.go:117] "RemoveContainer" containerID="92dd2e06edba76eba28215b6eab668acefde7b98db2adfddfd29ea06dd9b29b4" Dec 09 18:09:17 crc kubenswrapper[4954]: E1209 18:09:17.122842 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:09:20 crc kubenswrapper[4954]: E1209 18:09:20.122040 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:09:31 crc kubenswrapper[4954]: E1209 18:09:31.123648 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:09:31 crc kubenswrapper[4954]: E1209 18:09:31.124205 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:09:44 crc kubenswrapper[4954]: E1209 18:09:44.131779 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:09:46 crc kubenswrapper[4954]: E1209 18:09:46.122568 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:09:55 crc kubenswrapper[4954]: E1209 18:09:55.122163 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:09:57 crc kubenswrapper[4954]: E1209 18:09:57.124133 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:10:07 crc kubenswrapper[4954]: E1209 18:10:07.123180 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:10:12 crc kubenswrapper[4954]: E1209 18:10:12.124222 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:10:18 crc kubenswrapper[4954]: E1209 18:10:18.123294 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:10:24 crc kubenswrapper[4954]: E1209 18:10:24.129052 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:10:32 crc kubenswrapper[4954]: E1209 18:10:32.123511 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:10:35 crc kubenswrapper[4954]: E1209 18:10:35.122040 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:10:44 crc kubenswrapper[4954]: E1209 18:10:44.131021 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:10:49 crc kubenswrapper[4954]: E1209 18:10:49.122938 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:10:58 crc kubenswrapper[4954]: E1209 18:10:58.124519 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:11:04 crc kubenswrapper[4954]: E1209 18:11:04.131044 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:11:09 crc kubenswrapper[4954]: E1209 18:11:09.123275 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:11:17 crc kubenswrapper[4954]: E1209 18:11:17.122893 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.063083 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j4rwh"] Dec 09 18:11:20 crc kubenswrapper[4954]: E1209 18:11:20.064118 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerName="extract-content" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.064132 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerName="extract-content" Dec 09 18:11:20 crc kubenswrapper[4954]: E1209 18:11:20.064163 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerName="extract-utilities" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.064169 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerName="extract-utilities" Dec 09 18:11:20 crc kubenswrapper[4954]: E1209 18:11:20.064215 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerName="registry-server" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.064223 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerName="registry-server" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.064443 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="a17b20b0-c81f-42c5-96fb-19554d40f90b" containerName="registry-server" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.066342 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.076376 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j4rwh"] Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.238101 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxldc\" (UniqueName: \"kubernetes.io/projected/0ffead20-0858-4e11-8d58-2c85c1267cd6-kube-api-access-xxldc\") pod \"community-operators-j4rwh\" (UID: \"0ffead20-0858-4e11-8d58-2c85c1267cd6\") " pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.238243 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ffead20-0858-4e11-8d58-2c85c1267cd6-catalog-content\") pod \"community-operators-j4rwh\" (UID: \"0ffead20-0858-4e11-8d58-2c85c1267cd6\") " pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.238315 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ffead20-0858-4e11-8d58-2c85c1267cd6-utilities\") pod \"community-operators-j4rwh\" (UID: \"0ffead20-0858-4e11-8d58-2c85c1267cd6\") " pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.340006 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxldc\" (UniqueName: \"kubernetes.io/projected/0ffead20-0858-4e11-8d58-2c85c1267cd6-kube-api-access-xxldc\") pod \"community-operators-j4rwh\" (UID: \"0ffead20-0858-4e11-8d58-2c85c1267cd6\") " pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.340056 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ffead20-0858-4e11-8d58-2c85c1267cd6-catalog-content\") pod \"community-operators-j4rwh\" (UID: \"0ffead20-0858-4e11-8d58-2c85c1267cd6\") " pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.340117 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ffead20-0858-4e11-8d58-2c85c1267cd6-utilities\") pod \"community-operators-j4rwh\" (UID: \"0ffead20-0858-4e11-8d58-2c85c1267cd6\") " pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.340783 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ffead20-0858-4e11-8d58-2c85c1267cd6-utilities\") pod \"community-operators-j4rwh\" (UID: \"0ffead20-0858-4e11-8d58-2c85c1267cd6\") " pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.340996 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ffead20-0858-4e11-8d58-2c85c1267cd6-catalog-content\") pod \"community-operators-j4rwh\" (UID: \"0ffead20-0858-4e11-8d58-2c85c1267cd6\") " pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.368978 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxldc\" (UniqueName: \"kubernetes.io/projected/0ffead20-0858-4e11-8d58-2c85c1267cd6-kube-api-access-xxldc\") pod \"community-operators-j4rwh\" (UID: \"0ffead20-0858-4e11-8d58-2c85c1267cd6\") " pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:20 crc kubenswrapper[4954]: I1209 18:11:20.428868 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:21 crc kubenswrapper[4954]: I1209 18:11:21.122852 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j4rwh"] Dec 09 18:11:21 crc kubenswrapper[4954]: I1209 18:11:21.669834 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4rwh" event={"ID":"0ffead20-0858-4e11-8d58-2c85c1267cd6","Type":"ContainerStarted","Data":"78ea70c69bf1950a83d06adedb2e90efd3dce43d664387ce63138be32b1ae413"} Dec 09 18:11:22 crc kubenswrapper[4954]: I1209 18:11:22.680688 4954 generic.go:334] "Generic (PLEG): container finished" podID="0ffead20-0858-4e11-8d58-2c85c1267cd6" containerID="cce88d28ffd795dddb478c6d72b3e63a473043587ff7ce0d6495c5796626c047" exitCode=0 Dec 09 18:11:22 crc kubenswrapper[4954]: I1209 18:11:22.680746 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4rwh" event={"ID":"0ffead20-0858-4e11-8d58-2c85c1267cd6","Type":"ContainerDied","Data":"cce88d28ffd795dddb478c6d72b3e63a473043587ff7ce0d6495c5796626c047"} Dec 09 18:11:23 crc kubenswrapper[4954]: E1209 18:11:23.121068 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:11:27 crc kubenswrapper[4954]: I1209 18:11:27.747919 4954 generic.go:334] "Generic (PLEG): container finished" podID="0ffead20-0858-4e11-8d58-2c85c1267cd6" containerID="5db6c309ead8cd3cd81f4a961770e886d07937fa164028098cff21b9b0a3402b" exitCode=0 Dec 09 18:11:27 crc kubenswrapper[4954]: I1209 18:11:27.748016 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4rwh" event={"ID":"0ffead20-0858-4e11-8d58-2c85c1267cd6","Type":"ContainerDied","Data":"5db6c309ead8cd3cd81f4a961770e886d07937fa164028098cff21b9b0a3402b"} Dec 09 18:11:28 crc kubenswrapper[4954]: E1209 18:11:28.122833 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:11:29 crc kubenswrapper[4954]: I1209 18:11:29.778924 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4rwh" event={"ID":"0ffead20-0858-4e11-8d58-2c85c1267cd6","Type":"ContainerStarted","Data":"e32b14f50f101f4e91f582f18b9138cdb70b634cb022dc4f4bbdb927b80189b0"} Dec 09 18:11:29 crc kubenswrapper[4954]: I1209 18:11:29.800881 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j4rwh" podStartSLOduration=3.346295266 podStartE2EDuration="9.800862481s" podCreationTimestamp="2025-12-09 18:11:20 +0000 UTC" firstStartedPulling="2025-12-09 18:11:22.68292274 +0000 UTC m=+4479.071096560" lastFinishedPulling="2025-12-09 18:11:29.137489955 +0000 UTC m=+4485.525663775" observedRunningTime="2025-12-09 18:11:29.794629124 +0000 UTC m=+4486.182802944" watchObservedRunningTime="2025-12-09 18:11:29.800862481 +0000 UTC m=+4486.189036301" Dec 09 18:11:30 crc kubenswrapper[4954]: I1209 18:11:30.430265 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:30 crc kubenswrapper[4954]: I1209 18:11:30.430729 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:31 crc kubenswrapper[4954]: I1209 18:11:31.486187 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-j4rwh" podUID="0ffead20-0858-4e11-8d58-2c85c1267cd6" containerName="registry-server" probeResult="failure" output=< Dec 09 18:11:31 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 18:11:31 crc kubenswrapper[4954]: > Dec 09 18:11:34 crc kubenswrapper[4954]: E1209 18:11:34.130706 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:11:40 crc kubenswrapper[4954]: E1209 18:11:40.123384 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:11:40 crc kubenswrapper[4954]: I1209 18:11:40.479709 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:40 crc kubenswrapper[4954]: I1209 18:11:40.543612 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j4rwh" Dec 09 18:11:40 crc kubenswrapper[4954]: I1209 18:11:40.620038 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j4rwh"] Dec 09 18:11:40 crc kubenswrapper[4954]: I1209 18:11:40.721396 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2jsfq"] Dec 09 18:11:40 crc kubenswrapper[4954]: I1209 18:11:40.721724 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2jsfq" podUID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerName="registry-server" containerID="cri-o://e0d97a0477dd29addf0385ae26a2abeecbc332f4b210068fc1b6deb06096ea22" gracePeriod=2 Dec 09 18:11:40 crc kubenswrapper[4954]: I1209 18:11:40.894895 4954 generic.go:334] "Generic (PLEG): container finished" podID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerID="e0d97a0477dd29addf0385ae26a2abeecbc332f4b210068fc1b6deb06096ea22" exitCode=0 Dec 09 18:11:40 crc kubenswrapper[4954]: I1209 18:11:40.894957 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jsfq" event={"ID":"9ef7858f-b8fb-429b-886b-3671b1ab191e","Type":"ContainerDied","Data":"e0d97a0477dd29addf0385ae26a2abeecbc332f4b210068fc1b6deb06096ea22"} Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.228015 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2jsfq" Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.340253 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-catalog-content\") pod \"9ef7858f-b8fb-429b-886b-3671b1ab191e\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.340334 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-utilities\") pod \"9ef7858f-b8fb-429b-886b-3671b1ab191e\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.340431 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2ssq\" (UniqueName: \"kubernetes.io/projected/9ef7858f-b8fb-429b-886b-3671b1ab191e-kube-api-access-x2ssq\") pod \"9ef7858f-b8fb-429b-886b-3671b1ab191e\" (UID: \"9ef7858f-b8fb-429b-886b-3671b1ab191e\") " Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.341071 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-utilities" (OuterVolumeSpecName: "utilities") pod "9ef7858f-b8fb-429b-886b-3671b1ab191e" (UID: "9ef7858f-b8fb-429b-886b-3671b1ab191e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.341247 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.407551 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ef7858f-b8fb-429b-886b-3671b1ab191e" (UID: "9ef7858f-b8fb-429b-886b-3671b1ab191e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.442824 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef7858f-b8fb-429b-886b-3671b1ab191e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.828889 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ef7858f-b8fb-429b-886b-3671b1ab191e-kube-api-access-x2ssq" (OuterVolumeSpecName: "kube-api-access-x2ssq") pod "9ef7858f-b8fb-429b-886b-3671b1ab191e" (UID: "9ef7858f-b8fb-429b-886b-3671b1ab191e"). InnerVolumeSpecName "kube-api-access-x2ssq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.851586 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2ssq\" (UniqueName: \"kubernetes.io/projected/9ef7858f-b8fb-429b-886b-3671b1ab191e-kube-api-access-x2ssq\") on node \"crc\" DevicePath \"\"" Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.920303 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2jsfq" Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.923574 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jsfq" event={"ID":"9ef7858f-b8fb-429b-886b-3671b1ab191e","Type":"ContainerDied","Data":"5f8bd0e13d24ac263c41cef14987488d6f49d4a978a37e78c0452b4303cee0b5"} Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.923666 4954 scope.go:117] "RemoveContainer" containerID="e0d97a0477dd29addf0385ae26a2abeecbc332f4b210068fc1b6deb06096ea22" Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.966780 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2jsfq"] Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.979037 4954 scope.go:117] "RemoveContainer" containerID="157d27832ca42157552f6cb23c06586a2ea392db8ee12278155142ed7712aab2" Dec 09 18:11:41 crc kubenswrapper[4954]: I1209 18:11:41.979664 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2jsfq"] Dec 09 18:11:42 crc kubenswrapper[4954]: I1209 18:11:42.004626 4954 scope.go:117] "RemoveContainer" containerID="3993e73bf7b5be3995b3cdc2f27835f3d9dd3f579f1ac385fa589059ad803d72" Dec 09 18:11:42 crc kubenswrapper[4954]: I1209 18:11:42.135758 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ef7858f-b8fb-429b-886b-3671b1ab191e" path="/var/lib/kubelet/pods/9ef7858f-b8fb-429b-886b-3671b1ab191e/volumes" Dec 09 18:11:43 crc kubenswrapper[4954]: I1209 18:11:43.755532 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:11:43 crc kubenswrapper[4954]: I1209 18:11:43.755917 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:11:47 crc kubenswrapper[4954]: E1209 18:11:47.123602 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:11:55 crc kubenswrapper[4954]: E1209 18:11:55.122836 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:12:00 crc kubenswrapper[4954]: E1209 18:12:00.124075 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.522117 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kqqfm"] Dec 09 18:12:05 crc kubenswrapper[4954]: E1209 18:12:05.523221 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerName="extract-utilities" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.523236 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerName="extract-utilities" Dec 09 18:12:05 crc kubenswrapper[4954]: E1209 18:12:05.523247 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerName="registry-server" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.523253 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerName="registry-server" Dec 09 18:12:05 crc kubenswrapper[4954]: E1209 18:12:05.523274 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerName="extract-content" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.523280 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerName="extract-content" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.523535 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ef7858f-b8fb-429b-886b-3671b1ab191e" containerName="registry-server" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.525319 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.541495 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqqfm"] Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.656208 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-catalog-content\") pod \"certified-operators-kqqfm\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.656318 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z27z\" (UniqueName: \"kubernetes.io/projected/8ad25aff-7af3-47ee-9f0f-853a26af2a18-kube-api-access-8z27z\") pod \"certified-operators-kqqfm\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.656387 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-utilities\") pod \"certified-operators-kqqfm\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.758347 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z27z\" (UniqueName: \"kubernetes.io/projected/8ad25aff-7af3-47ee-9f0f-853a26af2a18-kube-api-access-8z27z\") pod \"certified-operators-kqqfm\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.758452 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-utilities\") pod \"certified-operators-kqqfm\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.758576 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-catalog-content\") pod \"certified-operators-kqqfm\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.759183 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-utilities\") pod \"certified-operators-kqqfm\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.759225 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-catalog-content\") pod \"certified-operators-kqqfm\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.778169 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z27z\" (UniqueName: \"kubernetes.io/projected/8ad25aff-7af3-47ee-9f0f-853a26af2a18-kube-api-access-8z27z\") pod \"certified-operators-kqqfm\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.859121 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:05 crc kubenswrapper[4954]: I1209 18:12:05.984185 4954 scope.go:117] "RemoveContainer" containerID="b11242f90e029d5732ffec39bd36e13d2fffd1148ee1721883f8dce9e0c54e16" Dec 09 18:12:06 crc kubenswrapper[4954]: I1209 18:12:06.049760 4954 scope.go:117] "RemoveContainer" containerID="34f1ebcaf5ec702324068294d3dd0e818a929bdbcb373aa5c47a0981a14e20a4" Dec 09 18:12:06 crc kubenswrapper[4954]: I1209 18:12:06.082117 4954 scope.go:117] "RemoveContainer" containerID="6ecfc7fb7cd54dbc6a9f41cd7c362c07a27cc115b06bb79adbb0878ab4640ca6" Dec 09 18:12:06 crc kubenswrapper[4954]: I1209 18:12:06.427804 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqqfm"] Dec 09 18:12:07 crc kubenswrapper[4954]: I1209 18:12:07.215819 4954 generic.go:334] "Generic (PLEG): container finished" podID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerID="74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a" exitCode=0 Dec 09 18:12:07 crc kubenswrapper[4954]: I1209 18:12:07.216153 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqqfm" event={"ID":"8ad25aff-7af3-47ee-9f0f-853a26af2a18","Type":"ContainerDied","Data":"74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a"} Dec 09 18:12:07 crc kubenswrapper[4954]: I1209 18:12:07.216207 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqqfm" event={"ID":"8ad25aff-7af3-47ee-9f0f-853a26af2a18","Type":"ContainerStarted","Data":"e584ec7a124038ee6c07da0230a335576697e5c4971aa84d229f10d506ebfaa5"} Dec 09 18:12:08 crc kubenswrapper[4954]: E1209 18:12:08.121998 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:12:09 crc kubenswrapper[4954]: I1209 18:12:09.236178 4954 generic.go:334] "Generic (PLEG): container finished" podID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerID="d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7" exitCode=0 Dec 09 18:12:09 crc kubenswrapper[4954]: I1209 18:12:09.236241 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqqfm" event={"ID":"8ad25aff-7af3-47ee-9f0f-853a26af2a18","Type":"ContainerDied","Data":"d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7"} Dec 09 18:12:11 crc kubenswrapper[4954]: I1209 18:12:11.257031 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqqfm" event={"ID":"8ad25aff-7af3-47ee-9f0f-853a26af2a18","Type":"ContainerStarted","Data":"21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448"} Dec 09 18:12:11 crc kubenswrapper[4954]: I1209 18:12:11.281514 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kqqfm" podStartSLOduration=3.422305901 podStartE2EDuration="6.281497315s" podCreationTimestamp="2025-12-09 18:12:05 +0000 UTC" firstStartedPulling="2025-12-09 18:12:07.21929058 +0000 UTC m=+4523.607464400" lastFinishedPulling="2025-12-09 18:12:10.078481994 +0000 UTC m=+4526.466655814" observedRunningTime="2025-12-09 18:12:11.278405337 +0000 UTC m=+4527.666579167" watchObservedRunningTime="2025-12-09 18:12:11.281497315 +0000 UTC m=+4527.669671135" Dec 09 18:12:13 crc kubenswrapper[4954]: I1209 18:12:13.756324 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:12:13 crc kubenswrapper[4954]: I1209 18:12:13.757576 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:12:14 crc kubenswrapper[4954]: E1209 18:12:14.129375 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:12:15 crc kubenswrapper[4954]: I1209 18:12:15.860480 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:15 crc kubenswrapper[4954]: I1209 18:12:15.860543 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:15 crc kubenswrapper[4954]: I1209 18:12:15.910557 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:16 crc kubenswrapper[4954]: I1209 18:12:16.370370 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:16 crc kubenswrapper[4954]: I1209 18:12:16.433952 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kqqfm"] Dec 09 18:12:18 crc kubenswrapper[4954]: I1209 18:12:18.339744 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kqqfm" podUID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerName="registry-server" containerID="cri-o://21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448" gracePeriod=2 Dec 09 18:12:18 crc kubenswrapper[4954]: I1209 18:12:18.846415 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.048048 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z27z\" (UniqueName: \"kubernetes.io/projected/8ad25aff-7af3-47ee-9f0f-853a26af2a18-kube-api-access-8z27z\") pod \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.048385 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-catalog-content\") pod \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.048434 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-utilities\") pod \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\" (UID: \"8ad25aff-7af3-47ee-9f0f-853a26af2a18\") " Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.049290 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-utilities" (OuterVolumeSpecName: "utilities") pod "8ad25aff-7af3-47ee-9f0f-853a26af2a18" (UID: "8ad25aff-7af3-47ee-9f0f-853a26af2a18"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.054411 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad25aff-7af3-47ee-9f0f-853a26af2a18-kube-api-access-8z27z" (OuterVolumeSpecName: "kube-api-access-8z27z") pod "8ad25aff-7af3-47ee-9f0f-853a26af2a18" (UID: "8ad25aff-7af3-47ee-9f0f-853a26af2a18"). InnerVolumeSpecName "kube-api-access-8z27z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.151482 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.151564 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z27z\" (UniqueName: \"kubernetes.io/projected/8ad25aff-7af3-47ee-9f0f-853a26af2a18-kube-api-access-8z27z\") on node \"crc\" DevicePath \"\"" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.299964 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ad25aff-7af3-47ee-9f0f-853a26af2a18" (UID: "8ad25aff-7af3-47ee-9f0f-853a26af2a18"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.355688 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ad25aff-7af3-47ee-9f0f-853a26af2a18-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.356734 4954 generic.go:334] "Generic (PLEG): container finished" podID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerID="21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448" exitCode=0 Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.356790 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqqfm" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.356789 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqqfm" event={"ID":"8ad25aff-7af3-47ee-9f0f-853a26af2a18","Type":"ContainerDied","Data":"21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448"} Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.356927 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqqfm" event={"ID":"8ad25aff-7af3-47ee-9f0f-853a26af2a18","Type":"ContainerDied","Data":"e584ec7a124038ee6c07da0230a335576697e5c4971aa84d229f10d506ebfaa5"} Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.356951 4954 scope.go:117] "RemoveContainer" containerID="21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.380349 4954 scope.go:117] "RemoveContainer" containerID="d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.409176 4954 scope.go:117] "RemoveContainer" containerID="74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.413710 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kqqfm"] Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.427418 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kqqfm"] Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.457158 4954 scope.go:117] "RemoveContainer" containerID="21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448" Dec 09 18:12:19 crc kubenswrapper[4954]: E1209 18:12:19.457640 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448\": container with ID starting with 21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448 not found: ID does not exist" containerID="21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.457673 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448"} err="failed to get container status \"21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448\": rpc error: code = NotFound desc = could not find container \"21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448\": container with ID starting with 21ddefdbd7f96cdc55b56b7382849fd30d8f01ee37d6223b0bfe7c4f4223a448 not found: ID does not exist" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.457695 4954 scope.go:117] "RemoveContainer" containerID="d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7" Dec 09 18:12:19 crc kubenswrapper[4954]: E1209 18:12:19.458200 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7\": container with ID starting with d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7 not found: ID does not exist" containerID="d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.458268 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7"} err="failed to get container status \"d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7\": rpc error: code = NotFound desc = could not find container \"d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7\": container with ID starting with d26c6e1ad2df95574c029c2000fb80a27d5edde8438e8c4e76e5932036c60ee7 not found: ID does not exist" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.458357 4954 scope.go:117] "RemoveContainer" containerID="74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a" Dec 09 18:12:19 crc kubenswrapper[4954]: E1209 18:12:19.458648 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a\": container with ID starting with 74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a not found: ID does not exist" containerID="74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a" Dec 09 18:12:19 crc kubenswrapper[4954]: I1209 18:12:19.458679 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a"} err="failed to get container status \"74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a\": rpc error: code = NotFound desc = could not find container \"74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a\": container with ID starting with 74bf9fd42bbafffa076aeb7f9c49c8c8a7f6036274c5587919bedcaccce6bb3a not found: ID does not exist" Dec 09 18:12:20 crc kubenswrapper[4954]: E1209 18:12:20.123186 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:12:20 crc kubenswrapper[4954]: I1209 18:12:20.137954 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" path="/var/lib/kubelet/pods/8ad25aff-7af3-47ee-9f0f-853a26af2a18/volumes" Dec 09 18:12:29 crc kubenswrapper[4954]: E1209 18:12:29.124033 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:12:33 crc kubenswrapper[4954]: E1209 18:12:33.125998 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:12:40 crc kubenswrapper[4954]: E1209 18:12:40.123261 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:12:43 crc kubenswrapper[4954]: I1209 18:12:43.755256 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:12:43 crc kubenswrapper[4954]: I1209 18:12:43.756085 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:12:43 crc kubenswrapper[4954]: I1209 18:12:43.756164 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 18:12:43 crc kubenswrapper[4954]: I1209 18:12:43.757038 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 18:12:43 crc kubenswrapper[4954]: I1209 18:12:43.757091 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" gracePeriod=600 Dec 09 18:12:44 crc kubenswrapper[4954]: E1209 18:12:44.132948 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:12:44 crc kubenswrapper[4954]: I1209 18:12:44.640266 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" exitCode=0 Dec 09 18:12:44 crc kubenswrapper[4954]: I1209 18:12:44.640609 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e"} Dec 09 18:12:44 crc kubenswrapper[4954]: I1209 18:12:44.640656 4954 scope.go:117] "RemoveContainer" containerID="40c79a44d3663bf94a20a1c5aea3afef5e17d945f82090de1b839c4a645080fd" Dec 09 18:12:44 crc kubenswrapper[4954]: E1209 18:12:44.648106 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:12:45 crc kubenswrapper[4954]: I1209 18:12:45.653217 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:12:45 crc kubenswrapper[4954]: E1209 18:12:45.653882 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:12:55 crc kubenswrapper[4954]: E1209 18:12:55.124003 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:12:58 crc kubenswrapper[4954]: E1209 18:12:58.123380 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:13:00 crc kubenswrapper[4954]: I1209 18:13:00.122310 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:13:00 crc kubenswrapper[4954]: E1209 18:13:00.122865 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:13:08 crc kubenswrapper[4954]: I1209 18:13:08.122487 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 18:13:08 crc kubenswrapper[4954]: E1209 18:13:08.239794 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:13:08 crc kubenswrapper[4954]: E1209 18:13:08.240174 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:13:08 crc kubenswrapper[4954]: E1209 18:13:08.240344 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:13:08 crc kubenswrapper[4954]: E1209 18:13:08.241561 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:13:10 crc kubenswrapper[4954]: E1209 18:13:10.122365 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:13:15 crc kubenswrapper[4954]: I1209 18:13:15.120569 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:13:15 crc kubenswrapper[4954]: E1209 18:13:15.121391 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:13:23 crc kubenswrapper[4954]: E1209 18:13:23.125351 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:13:24 crc kubenswrapper[4954]: E1209 18:13:24.239634 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:13:24 crc kubenswrapper[4954]: E1209 18:13:24.240258 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:13:24 crc kubenswrapper[4954]: E1209 18:13:24.240511 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:13:24 crc kubenswrapper[4954]: E1209 18:13:24.241868 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:13:26 crc kubenswrapper[4954]: I1209 18:13:26.120258 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:13:26 crc kubenswrapper[4954]: E1209 18:13:26.120793 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:13:36 crc kubenswrapper[4954]: E1209 18:13:36.122466 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:13:37 crc kubenswrapper[4954]: E1209 18:13:37.123527 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:13:41 crc kubenswrapper[4954]: I1209 18:13:41.120802 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:13:41 crc kubenswrapper[4954]: E1209 18:13:41.121633 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:13:48 crc kubenswrapper[4954]: E1209 18:13:48.122320 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:13:49 crc kubenswrapper[4954]: E1209 18:13:49.122147 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:13:53 crc kubenswrapper[4954]: I1209 18:13:53.120253 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:13:53 crc kubenswrapper[4954]: E1209 18:13:53.121133 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:14:00 crc kubenswrapper[4954]: E1209 18:14:00.122887 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:14:02 crc kubenswrapper[4954]: E1209 18:14:02.122612 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:14:07 crc kubenswrapper[4954]: I1209 18:14:07.121136 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:14:07 crc kubenswrapper[4954]: E1209 18:14:07.122385 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:14:11 crc kubenswrapper[4954]: E1209 18:14:11.123649 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:14:16 crc kubenswrapper[4954]: E1209 18:14:16.122788 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:14:19 crc kubenswrapper[4954]: I1209 18:14:19.121630 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:14:19 crc kubenswrapper[4954]: E1209 18:14:19.122796 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:14:22 crc kubenswrapper[4954]: E1209 18:14:22.123816 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:14:28 crc kubenswrapper[4954]: E1209 18:14:28.124926 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:14:34 crc kubenswrapper[4954]: I1209 18:14:34.127806 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:14:34 crc kubenswrapper[4954]: E1209 18:14:34.128610 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:14:35 crc kubenswrapper[4954]: E1209 18:14:35.125036 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:14:42 crc kubenswrapper[4954]: E1209 18:14:42.123219 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:14:46 crc kubenswrapper[4954]: I1209 18:14:46.120510 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:14:46 crc kubenswrapper[4954]: E1209 18:14:46.121120 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:14:48 crc kubenswrapper[4954]: E1209 18:14:48.122578 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:14:57 crc kubenswrapper[4954]: E1209 18:14:57.124433 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:14:58 crc kubenswrapper[4954]: I1209 18:14:58.121307 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:14:58 crc kubenswrapper[4954]: E1209 18:14:58.121972 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.730516 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt"] Dec 09 18:15:00 crc kubenswrapper[4954]: E1209 18:15:00.731649 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerName="extract-utilities" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.731667 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerName="extract-utilities" Dec 09 18:15:00 crc kubenswrapper[4954]: E1209 18:15:00.731679 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerName="extract-content" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.731687 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerName="extract-content" Dec 09 18:15:00 crc kubenswrapper[4954]: E1209 18:15:00.731699 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerName="registry-server" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.731708 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerName="registry-server" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.731990 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad25aff-7af3-47ee-9f0f-853a26af2a18" containerName="registry-server" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.732967 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.734980 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.735202 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.746923 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt"] Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.842572 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6258145-2397-4644-b14f-e5602942c6c8-secret-volume\") pod \"collect-profiles-29421735-hwglt\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.842700 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6258145-2397-4644-b14f-e5602942c6c8-config-volume\") pod \"collect-profiles-29421735-hwglt\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.842796 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr44v\" (UniqueName: \"kubernetes.io/projected/e6258145-2397-4644-b14f-e5602942c6c8-kube-api-access-xr44v\") pod \"collect-profiles-29421735-hwglt\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.944776 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6258145-2397-4644-b14f-e5602942c6c8-config-volume\") pod \"collect-profiles-29421735-hwglt\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.944880 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr44v\" (UniqueName: \"kubernetes.io/projected/e6258145-2397-4644-b14f-e5602942c6c8-kube-api-access-xr44v\") pod \"collect-profiles-29421735-hwglt\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.944999 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6258145-2397-4644-b14f-e5602942c6c8-secret-volume\") pod \"collect-profiles-29421735-hwglt\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.946167 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6258145-2397-4644-b14f-e5602942c6c8-config-volume\") pod \"collect-profiles-29421735-hwglt\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.958582 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6258145-2397-4644-b14f-e5602942c6c8-secret-volume\") pod \"collect-profiles-29421735-hwglt\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:00 crc kubenswrapper[4954]: I1209 18:15:00.961390 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr44v\" (UniqueName: \"kubernetes.io/projected/e6258145-2397-4644-b14f-e5602942c6c8-kube-api-access-xr44v\") pod \"collect-profiles-29421735-hwglt\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:01 crc kubenswrapper[4954]: I1209 18:15:01.056960 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:01 crc kubenswrapper[4954]: I1209 18:15:01.499903 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt"] Dec 09 18:15:01 crc kubenswrapper[4954]: I1209 18:15:01.692291 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" event={"ID":"e6258145-2397-4644-b14f-e5602942c6c8","Type":"ContainerStarted","Data":"dc25e38f2df078df90767dbfa665099e060b8567e63cdefd67693c1262b2eddb"} Dec 09 18:15:01 crc kubenswrapper[4954]: I1209 18:15:01.692679 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" event={"ID":"e6258145-2397-4644-b14f-e5602942c6c8","Type":"ContainerStarted","Data":"527fdd6d71342840c82aaf6b68f092f2c88cef3f4298108008a36206c6aaa12f"} Dec 09 18:15:01 crc kubenswrapper[4954]: I1209 18:15:01.710850 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" podStartSLOduration=1.710827477 podStartE2EDuration="1.710827477s" podCreationTimestamp="2025-12-09 18:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 18:15:01.70649207 +0000 UTC m=+4698.094665910" watchObservedRunningTime="2025-12-09 18:15:01.710827477 +0000 UTC m=+4698.099001297" Dec 09 18:15:02 crc kubenswrapper[4954]: E1209 18:15:02.122171 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:15:02 crc kubenswrapper[4954]: I1209 18:15:02.707978 4954 generic.go:334] "Generic (PLEG): container finished" podID="e6258145-2397-4644-b14f-e5602942c6c8" containerID="dc25e38f2df078df90767dbfa665099e060b8567e63cdefd67693c1262b2eddb" exitCode=0 Dec 09 18:15:02 crc kubenswrapper[4954]: I1209 18:15:02.708019 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" event={"ID":"e6258145-2397-4644-b14f-e5602942c6c8","Type":"ContainerDied","Data":"dc25e38f2df078df90767dbfa665099e060b8567e63cdefd67693c1262b2eddb"} Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.043008 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.117862 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6258145-2397-4644-b14f-e5602942c6c8-config-volume\") pod \"e6258145-2397-4644-b14f-e5602942c6c8\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.118119 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6258145-2397-4644-b14f-e5602942c6c8-secret-volume\") pod \"e6258145-2397-4644-b14f-e5602942c6c8\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.118912 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr44v\" (UniqueName: \"kubernetes.io/projected/e6258145-2397-4644-b14f-e5602942c6c8-kube-api-access-xr44v\") pod \"e6258145-2397-4644-b14f-e5602942c6c8\" (UID: \"e6258145-2397-4644-b14f-e5602942c6c8\") " Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.119360 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6258145-2397-4644-b14f-e5602942c6c8-config-volume" (OuterVolumeSpecName: "config-volume") pod "e6258145-2397-4644-b14f-e5602942c6c8" (UID: "e6258145-2397-4644-b14f-e5602942c6c8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.120580 4954 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6258145-2397-4644-b14f-e5602942c6c8-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.124659 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6258145-2397-4644-b14f-e5602942c6c8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e6258145-2397-4644-b14f-e5602942c6c8" (UID: "e6258145-2397-4644-b14f-e5602942c6c8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.125288 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6258145-2397-4644-b14f-e5602942c6c8-kube-api-access-xr44v" (OuterVolumeSpecName: "kube-api-access-xr44v") pod "e6258145-2397-4644-b14f-e5602942c6c8" (UID: "e6258145-2397-4644-b14f-e5602942c6c8"). InnerVolumeSpecName "kube-api-access-xr44v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.223417 4954 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6258145-2397-4644-b14f-e5602942c6c8-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.223456 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr44v\" (UniqueName: \"kubernetes.io/projected/e6258145-2397-4644-b14f-e5602942c6c8-kube-api-access-xr44v\") on node \"crc\" DevicePath \"\"" Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.727986 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" event={"ID":"e6258145-2397-4644-b14f-e5602942c6c8","Type":"ContainerDied","Data":"527fdd6d71342840c82aaf6b68f092f2c88cef3f4298108008a36206c6aaa12f"} Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.728038 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="527fdd6d71342840c82aaf6b68f092f2c88cef3f4298108008a36206c6aaa12f" Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.728082 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421735-hwglt" Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.780902 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s"] Dec 09 18:15:04 crc kubenswrapper[4954]: I1209 18:15:04.791486 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421690-dkg8s"] Dec 09 18:15:06 crc kubenswrapper[4954]: I1209 18:15:06.132234 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f789cd82-95d6-4192-b4cf-586ff93babcf" path="/var/lib/kubelet/pods/f789cd82-95d6-4192-b4cf-586ff93babcf/volumes" Dec 09 18:15:06 crc kubenswrapper[4954]: I1209 18:15:06.274900 4954 scope.go:117] "RemoveContainer" containerID="0673e594c40666bba9c0a3d7507149040669bdd1282cf18fa4e895d6d81b87b5" Dec 09 18:15:10 crc kubenswrapper[4954]: E1209 18:15:10.123733 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:15:12 crc kubenswrapper[4954]: I1209 18:15:12.121054 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:15:12 crc kubenswrapper[4954]: E1209 18:15:12.121667 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:15:16 crc kubenswrapper[4954]: E1209 18:15:16.122464 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:15:21 crc kubenswrapper[4954]: E1209 18:15:21.123826 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:15:26 crc kubenswrapper[4954]: I1209 18:15:26.121307 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:15:26 crc kubenswrapper[4954]: E1209 18:15:26.122389 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:15:31 crc kubenswrapper[4954]: E1209 18:15:31.123281 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:15:36 crc kubenswrapper[4954]: E1209 18:15:36.122422 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:15:38 crc kubenswrapper[4954]: I1209 18:15:38.120165 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:15:38 crc kubenswrapper[4954]: E1209 18:15:38.120846 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:15:46 crc kubenswrapper[4954]: E1209 18:15:46.122491 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:15:49 crc kubenswrapper[4954]: I1209 18:15:49.121042 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:15:49 crc kubenswrapper[4954]: E1209 18:15:49.121696 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:15:50 crc kubenswrapper[4954]: E1209 18:15:50.122913 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:16:01 crc kubenswrapper[4954]: E1209 18:16:01.122707 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:16:02 crc kubenswrapper[4954]: I1209 18:16:02.275308 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:16:02 crc kubenswrapper[4954]: E1209 18:16:02.278457 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:16:03 crc kubenswrapper[4954]: E1209 18:16:03.122404 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:16:09 crc kubenswrapper[4954]: I1209 18:16:09.999495 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vd6xx"] Dec 09 18:16:10 crc kubenswrapper[4954]: E1209 18:16:10.001797 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6258145-2397-4644-b14f-e5602942c6c8" containerName="collect-profiles" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.001930 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6258145-2397-4644-b14f-e5602942c6c8" containerName="collect-profiles" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.004011 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6258145-2397-4644-b14f-e5602942c6c8" containerName="collect-profiles" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.006624 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.031465 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vd6xx"] Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.102076 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-utilities\") pod \"redhat-operators-vd6xx\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.102158 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wchzv\" (UniqueName: \"kubernetes.io/projected/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-kube-api-access-wchzv\") pod \"redhat-operators-vd6xx\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.102238 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-catalog-content\") pod \"redhat-operators-vd6xx\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.204181 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-utilities\") pod \"redhat-operators-vd6xx\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.204270 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wchzv\" (UniqueName: \"kubernetes.io/projected/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-kube-api-access-wchzv\") pod \"redhat-operators-vd6xx\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.204348 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-catalog-content\") pod \"redhat-operators-vd6xx\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.205043 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-utilities\") pod \"redhat-operators-vd6xx\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.205087 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-catalog-content\") pod \"redhat-operators-vd6xx\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.225267 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wchzv\" (UniqueName: \"kubernetes.io/projected/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-kube-api-access-wchzv\") pod \"redhat-operators-vd6xx\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.337218 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:10 crc kubenswrapper[4954]: I1209 18:16:10.883248 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vd6xx"] Dec 09 18:16:11 crc kubenswrapper[4954]: I1209 18:16:11.452243 4954 generic.go:334] "Generic (PLEG): container finished" podID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerID="88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a" exitCode=0 Dec 09 18:16:11 crc kubenswrapper[4954]: I1209 18:16:11.452350 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vd6xx" event={"ID":"7067f5f9-aea5-46f4-bce6-1e1bebe2307c","Type":"ContainerDied","Data":"88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a"} Dec 09 18:16:11 crc kubenswrapper[4954]: I1209 18:16:11.452527 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vd6xx" event={"ID":"7067f5f9-aea5-46f4-bce6-1e1bebe2307c","Type":"ContainerStarted","Data":"5988dc19c91a6f8d1c3469cfaa78860630f43942a8c29e441fe7694f406e5212"} Dec 09 18:16:13 crc kubenswrapper[4954]: I1209 18:16:13.559699 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vd6xx" event={"ID":"7067f5f9-aea5-46f4-bce6-1e1bebe2307c","Type":"ContainerStarted","Data":"96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893"} Dec 09 18:16:14 crc kubenswrapper[4954]: I1209 18:16:14.134618 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:16:14 crc kubenswrapper[4954]: E1209 18:16:14.134916 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:16:15 crc kubenswrapper[4954]: E1209 18:16:15.122368 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:16:15 crc kubenswrapper[4954]: I1209 18:16:15.578925 4954 generic.go:334] "Generic (PLEG): container finished" podID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerID="96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893" exitCode=0 Dec 09 18:16:15 crc kubenswrapper[4954]: I1209 18:16:15.578965 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vd6xx" event={"ID":"7067f5f9-aea5-46f4-bce6-1e1bebe2307c","Type":"ContainerDied","Data":"96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893"} Dec 09 18:16:16 crc kubenswrapper[4954]: E1209 18:16:16.121733 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:16:16 crc kubenswrapper[4954]: I1209 18:16:16.590571 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vd6xx" event={"ID":"7067f5f9-aea5-46f4-bce6-1e1bebe2307c","Type":"ContainerStarted","Data":"e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35"} Dec 09 18:16:16 crc kubenswrapper[4954]: I1209 18:16:16.622523 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vd6xx" podStartSLOduration=3.015030122 podStartE2EDuration="7.622499987s" podCreationTimestamp="2025-12-09 18:16:09 +0000 UTC" firstStartedPulling="2025-12-09 18:16:11.45453111 +0000 UTC m=+4767.842704930" lastFinishedPulling="2025-12-09 18:16:16.062000975 +0000 UTC m=+4772.450174795" observedRunningTime="2025-12-09 18:16:16.613010147 +0000 UTC m=+4773.001183967" watchObservedRunningTime="2025-12-09 18:16:16.622499987 +0000 UTC m=+4773.010673827" Dec 09 18:16:20 crc kubenswrapper[4954]: I1209 18:16:20.337813 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:20 crc kubenswrapper[4954]: I1209 18:16:20.338446 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:21 crc kubenswrapper[4954]: I1209 18:16:21.392199 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vd6xx" podUID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerName="registry-server" probeResult="failure" output=< Dec 09 18:16:21 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 18:16:21 crc kubenswrapper[4954]: > Dec 09 18:16:25 crc kubenswrapper[4954]: I1209 18:16:25.121039 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:16:25 crc kubenswrapper[4954]: E1209 18:16:25.121935 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:16:28 crc kubenswrapper[4954]: E1209 18:16:28.123861 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:16:29 crc kubenswrapper[4954]: E1209 18:16:29.122494 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:16:30 crc kubenswrapper[4954]: I1209 18:16:30.438889 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:30 crc kubenswrapper[4954]: I1209 18:16:30.500998 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:30 crc kubenswrapper[4954]: I1209 18:16:30.675055 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vd6xx"] Dec 09 18:16:31 crc kubenswrapper[4954]: I1209 18:16:31.998879 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vd6xx" podUID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerName="registry-server" containerID="cri-o://e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35" gracePeriod=2 Dec 09 18:16:32 crc kubenswrapper[4954]: I1209 18:16:32.528399 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:32 crc kubenswrapper[4954]: I1209 18:16:32.683892 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-catalog-content\") pod \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " Dec 09 18:16:32 crc kubenswrapper[4954]: I1209 18:16:32.684212 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-utilities\") pod \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " Dec 09 18:16:32 crc kubenswrapper[4954]: I1209 18:16:32.684262 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wchzv\" (UniqueName: \"kubernetes.io/projected/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-kube-api-access-wchzv\") pod \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\" (UID: \"7067f5f9-aea5-46f4-bce6-1e1bebe2307c\") " Dec 09 18:16:32 crc kubenswrapper[4954]: I1209 18:16:32.684962 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-utilities" (OuterVolumeSpecName: "utilities") pod "7067f5f9-aea5-46f4-bce6-1e1bebe2307c" (UID: "7067f5f9-aea5-46f4-bce6-1e1bebe2307c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:16:32 crc kubenswrapper[4954]: I1209 18:16:32.692890 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-kube-api-access-wchzv" (OuterVolumeSpecName: "kube-api-access-wchzv") pod "7067f5f9-aea5-46f4-bce6-1e1bebe2307c" (UID: "7067f5f9-aea5-46f4-bce6-1e1bebe2307c"). InnerVolumeSpecName "kube-api-access-wchzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:16:32 crc kubenswrapper[4954]: I1209 18:16:32.786887 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:16:32 crc kubenswrapper[4954]: I1209 18:16:32.786927 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wchzv\" (UniqueName: \"kubernetes.io/projected/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-kube-api-access-wchzv\") on node \"crc\" DevicePath \"\"" Dec 09 18:16:32 crc kubenswrapper[4954]: I1209 18:16:32.801326 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7067f5f9-aea5-46f4-bce6-1e1bebe2307c" (UID: "7067f5f9-aea5-46f4-bce6-1e1bebe2307c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:16:32 crc kubenswrapper[4954]: I1209 18:16:32.888788 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7067f5f9-aea5-46f4-bce6-1e1bebe2307c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.009566 4954 generic.go:334] "Generic (PLEG): container finished" podID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerID="e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35" exitCode=0 Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.009620 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vd6xx" event={"ID":"7067f5f9-aea5-46f4-bce6-1e1bebe2307c","Type":"ContainerDied","Data":"e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35"} Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.009649 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vd6xx" event={"ID":"7067f5f9-aea5-46f4-bce6-1e1bebe2307c","Type":"ContainerDied","Data":"5988dc19c91a6f8d1c3469cfaa78860630f43942a8c29e441fe7694f406e5212"} Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.009668 4954 scope.go:117] "RemoveContainer" containerID="e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35" Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.009696 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vd6xx" Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.034297 4954 scope.go:117] "RemoveContainer" containerID="96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893" Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.044561 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vd6xx"] Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.056180 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vd6xx"] Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.067387 4954 scope.go:117] "RemoveContainer" containerID="88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a" Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.195782 4954 scope.go:117] "RemoveContainer" containerID="e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35" Dec 09 18:16:33 crc kubenswrapper[4954]: E1209 18:16:33.209026 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35\": container with ID starting with e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35 not found: ID does not exist" containerID="e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35" Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.209092 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35"} err="failed to get container status \"e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35\": rpc error: code = NotFound desc = could not find container \"e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35\": container with ID starting with e63aaf1575e090fa43c989a0886c9beba383c1604f3f44253e8b45440c027f35 not found: ID does not exist" Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.209127 4954 scope.go:117] "RemoveContainer" containerID="96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893" Dec 09 18:16:33 crc kubenswrapper[4954]: E1209 18:16:33.220801 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893\": container with ID starting with 96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893 not found: ID does not exist" containerID="96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893" Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.220868 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893"} err="failed to get container status \"96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893\": rpc error: code = NotFound desc = could not find container \"96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893\": container with ID starting with 96f72ca165bbf835c581b13b16bdcde15e116fc89bfbd63aaf743e9117f3c893 not found: ID does not exist" Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.220907 4954 scope.go:117] "RemoveContainer" containerID="88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a" Dec 09 18:16:33 crc kubenswrapper[4954]: E1209 18:16:33.224759 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a\": container with ID starting with 88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a not found: ID does not exist" containerID="88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a" Dec 09 18:16:33 crc kubenswrapper[4954]: I1209 18:16:33.224806 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a"} err="failed to get container status \"88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a\": rpc error: code = NotFound desc = could not find container \"88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a\": container with ID starting with 88750c874fdb046b6f183ec774df8542d3d7f1065f54e63df27d7c092246708a not found: ID does not exist" Dec 09 18:16:34 crc kubenswrapper[4954]: I1209 18:16:34.135039 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" path="/var/lib/kubelet/pods/7067f5f9-aea5-46f4-bce6-1e1bebe2307c/volumes" Dec 09 18:16:38 crc kubenswrapper[4954]: I1209 18:16:38.122039 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:16:38 crc kubenswrapper[4954]: E1209 18:16:38.122875 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:16:39 crc kubenswrapper[4954]: E1209 18:16:39.123887 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:16:42 crc kubenswrapper[4954]: E1209 18:16:42.124944 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:16:42 crc kubenswrapper[4954]: I1209 18:16:42.485076 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-547f8767c7-hv854" podUID="a6635c8e-d9a5-4034-ab5c-7fe96dc10c10" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 09 18:16:50 crc kubenswrapper[4954]: I1209 18:16:50.120437 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:16:50 crc kubenswrapper[4954]: E1209 18:16:50.121385 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:16:51 crc kubenswrapper[4954]: E1209 18:16:51.123006 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:16:54 crc kubenswrapper[4954]: E1209 18:16:54.128800 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:17:03 crc kubenswrapper[4954]: E1209 18:17:03.124061 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:17:04 crc kubenswrapper[4954]: I1209 18:17:04.129918 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:17:04 crc kubenswrapper[4954]: E1209 18:17:04.130921 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:17:08 crc kubenswrapper[4954]: E1209 18:17:08.122443 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:17:15 crc kubenswrapper[4954]: E1209 18:17:15.124021 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:17:16 crc kubenswrapper[4954]: I1209 18:17:16.120315 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:17:16 crc kubenswrapper[4954]: E1209 18:17:16.120904 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:17:19 crc kubenswrapper[4954]: E1209 18:17:19.122876 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:17:28 crc kubenswrapper[4954]: I1209 18:17:28.120526 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:17:28 crc kubenswrapper[4954]: E1209 18:17:28.121405 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:17:29 crc kubenswrapper[4954]: E1209 18:17:29.122019 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:17:31 crc kubenswrapper[4954]: E1209 18:17:31.122341 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:17:41 crc kubenswrapper[4954]: I1209 18:17:41.120661 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:17:41 crc kubenswrapper[4954]: E1209 18:17:41.121505 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:17:43 crc kubenswrapper[4954]: E1209 18:17:43.123229 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:17:45 crc kubenswrapper[4954]: E1209 18:17:45.122515 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:17:53 crc kubenswrapper[4954]: I1209 18:17:53.121885 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:17:53 crc kubenswrapper[4954]: I1209 18:17:53.399810 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"f20ebcb5ad0c52185fb6ec22cec610c627bdbfc1f3c3a1e727e08a6d7a900c5b"} Dec 09 18:17:54 crc kubenswrapper[4954]: E1209 18:17:54.128857 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:17:57 crc kubenswrapper[4954]: E1209 18:17:57.122539 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:18:05 crc kubenswrapper[4954]: E1209 18:18:05.123952 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:18:10 crc kubenswrapper[4954]: E1209 18:18:10.123560 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:18:16 crc kubenswrapper[4954]: I1209 18:18:16.123640 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 18:18:16 crc kubenswrapper[4954]: E1209 18:18:16.248280 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:18:16 crc kubenswrapper[4954]: E1209 18:18:16.248359 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:18:16 crc kubenswrapper[4954]: E1209 18:18:16.248668 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:18:16 crc kubenswrapper[4954]: E1209 18:18:16.249975 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.098583 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lrs4n"] Dec 09 18:18:19 crc kubenswrapper[4954]: E1209 18:18:19.099708 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerName="registry-server" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.099728 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerName="registry-server" Dec 09 18:18:19 crc kubenswrapper[4954]: E1209 18:18:19.099759 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerName="extract-utilities" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.099769 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerName="extract-utilities" Dec 09 18:18:19 crc kubenswrapper[4954]: E1209 18:18:19.099785 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerName="extract-content" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.099791 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerName="extract-content" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.100019 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="7067f5f9-aea5-46f4-bce6-1e1bebe2307c" containerName="registry-server" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.101873 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.112622 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lrs4n"] Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.263170 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-catalog-content\") pod \"redhat-marketplace-lrs4n\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.263716 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gjp9\" (UniqueName: \"kubernetes.io/projected/3efb54d9-560f-415c-b533-2236952dcca4-kube-api-access-5gjp9\") pod \"redhat-marketplace-lrs4n\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.263752 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-utilities\") pod \"redhat-marketplace-lrs4n\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.367998 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gjp9\" (UniqueName: \"kubernetes.io/projected/3efb54d9-560f-415c-b533-2236952dcca4-kube-api-access-5gjp9\") pod \"redhat-marketplace-lrs4n\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.368376 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-utilities\") pod \"redhat-marketplace-lrs4n\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.368777 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-catalog-content\") pod \"redhat-marketplace-lrs4n\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.369060 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-utilities\") pod \"redhat-marketplace-lrs4n\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.369405 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-catalog-content\") pod \"redhat-marketplace-lrs4n\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.389351 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gjp9\" (UniqueName: \"kubernetes.io/projected/3efb54d9-560f-415c-b533-2236952dcca4-kube-api-access-5gjp9\") pod \"redhat-marketplace-lrs4n\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:19 crc kubenswrapper[4954]: I1209 18:18:19.427287 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:20 crc kubenswrapper[4954]: I1209 18:18:20.006369 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lrs4n"] Dec 09 18:18:20 crc kubenswrapper[4954]: I1209 18:18:20.746270 4954 generic.go:334] "Generic (PLEG): container finished" podID="3efb54d9-560f-415c-b533-2236952dcca4" containerID="3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6" exitCode=0 Dec 09 18:18:20 crc kubenswrapper[4954]: I1209 18:18:20.746384 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrs4n" event={"ID":"3efb54d9-560f-415c-b533-2236952dcca4","Type":"ContainerDied","Data":"3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6"} Dec 09 18:18:20 crc kubenswrapper[4954]: I1209 18:18:20.746954 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrs4n" event={"ID":"3efb54d9-560f-415c-b533-2236952dcca4","Type":"ContainerStarted","Data":"2598ae583533a192e0b0f7383b7da34e23c31b8c407e543262c44df25cc98825"} Dec 09 18:18:21 crc kubenswrapper[4954]: E1209 18:18:21.121531 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:18:22 crc kubenswrapper[4954]: I1209 18:18:22.767193 4954 generic.go:334] "Generic (PLEG): container finished" podID="3efb54d9-560f-415c-b533-2236952dcca4" containerID="16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7" exitCode=0 Dec 09 18:18:22 crc kubenswrapper[4954]: I1209 18:18:22.767258 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrs4n" event={"ID":"3efb54d9-560f-415c-b533-2236952dcca4","Type":"ContainerDied","Data":"16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7"} Dec 09 18:18:23 crc kubenswrapper[4954]: I1209 18:18:23.778214 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrs4n" event={"ID":"3efb54d9-560f-415c-b533-2236952dcca4","Type":"ContainerStarted","Data":"f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9"} Dec 09 18:18:23 crc kubenswrapper[4954]: I1209 18:18:23.820674 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lrs4n" podStartSLOduration=2.28682185 podStartE2EDuration="4.820651803s" podCreationTimestamp="2025-12-09 18:18:19 +0000 UTC" firstStartedPulling="2025-12-09 18:18:20.748700391 +0000 UTC m=+4897.136874211" lastFinishedPulling="2025-12-09 18:18:23.282530344 +0000 UTC m=+4899.670704164" observedRunningTime="2025-12-09 18:18:23.81679325 +0000 UTC m=+4900.204967060" watchObservedRunningTime="2025-12-09 18:18:23.820651803 +0000 UTC m=+4900.208825623" Dec 09 18:18:29 crc kubenswrapper[4954]: E1209 18:18:29.123827 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:18:29 crc kubenswrapper[4954]: I1209 18:18:29.429061 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:29 crc kubenswrapper[4954]: I1209 18:18:29.429132 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:29 crc kubenswrapper[4954]: I1209 18:18:29.478221 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:29 crc kubenswrapper[4954]: I1209 18:18:29.908973 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.092258 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lrs4n"] Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.093230 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lrs4n" podUID="3efb54d9-560f-415c-b533-2236952dcca4" containerName="registry-server" containerID="cri-o://f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9" gracePeriod=2 Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.634678 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.683114 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-catalog-content\") pod \"3efb54d9-560f-415c-b533-2236952dcca4\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.683380 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gjp9\" (UniqueName: \"kubernetes.io/projected/3efb54d9-560f-415c-b533-2236952dcca4-kube-api-access-5gjp9\") pod \"3efb54d9-560f-415c-b533-2236952dcca4\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.683410 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-utilities\") pod \"3efb54d9-560f-415c-b533-2236952dcca4\" (UID: \"3efb54d9-560f-415c-b533-2236952dcca4\") " Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.684192 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-utilities" (OuterVolumeSpecName: "utilities") pod "3efb54d9-560f-415c-b533-2236952dcca4" (UID: "3efb54d9-560f-415c-b533-2236952dcca4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.692399 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3efb54d9-560f-415c-b533-2236952dcca4-kube-api-access-5gjp9" (OuterVolumeSpecName: "kube-api-access-5gjp9") pod "3efb54d9-560f-415c-b533-2236952dcca4" (UID: "3efb54d9-560f-415c-b533-2236952dcca4"). InnerVolumeSpecName "kube-api-access-5gjp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.710120 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3efb54d9-560f-415c-b533-2236952dcca4" (UID: "3efb54d9-560f-415c-b533-2236952dcca4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.785580 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gjp9\" (UniqueName: \"kubernetes.io/projected/3efb54d9-560f-415c-b533-2236952dcca4-kube-api-access-5gjp9\") on node \"crc\" DevicePath \"\"" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.785627 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.785638 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efb54d9-560f-415c-b533-2236952dcca4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.897962 4954 generic.go:334] "Generic (PLEG): container finished" podID="3efb54d9-560f-415c-b533-2236952dcca4" containerID="f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9" exitCode=0 Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.898006 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrs4n" event={"ID":"3efb54d9-560f-415c-b533-2236952dcca4","Type":"ContainerDied","Data":"f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9"} Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.898033 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lrs4n" event={"ID":"3efb54d9-560f-415c-b533-2236952dcca4","Type":"ContainerDied","Data":"2598ae583533a192e0b0f7383b7da34e23c31b8c407e543262c44df25cc98825"} Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.898051 4954 scope.go:117] "RemoveContainer" containerID="f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.898211 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lrs4n" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.935224 4954 scope.go:117] "RemoveContainer" containerID="16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.942390 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lrs4n"] Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.954270 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lrs4n"] Dec 09 18:18:33 crc kubenswrapper[4954]: E1209 18:18:33.960443 4954 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3efb54d9_560f_415c_b533_2236952dcca4.slice/crio-2598ae583533a192e0b0f7383b7da34e23c31b8c407e543262c44df25cc98825\": RecentStats: unable to find data in memory cache]" Dec 09 18:18:33 crc kubenswrapper[4954]: I1209 18:18:33.967822 4954 scope.go:117] "RemoveContainer" containerID="3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6" Dec 09 18:18:34 crc kubenswrapper[4954]: I1209 18:18:34.021213 4954 scope.go:117] "RemoveContainer" containerID="f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9" Dec 09 18:18:34 crc kubenswrapper[4954]: E1209 18:18:34.024994 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9\": container with ID starting with f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9 not found: ID does not exist" containerID="f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9" Dec 09 18:18:34 crc kubenswrapper[4954]: I1209 18:18:34.025040 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9"} err="failed to get container status \"f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9\": rpc error: code = NotFound desc = could not find container \"f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9\": container with ID starting with f12f9dc8b8aa14c0189243c4ccb8df73f039212c0d9b2ca47be44ccdadbfffc9 not found: ID does not exist" Dec 09 18:18:34 crc kubenswrapper[4954]: I1209 18:18:34.025072 4954 scope.go:117] "RemoveContainer" containerID="16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7" Dec 09 18:18:34 crc kubenswrapper[4954]: E1209 18:18:34.025506 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7\": container with ID starting with 16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7 not found: ID does not exist" containerID="16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7" Dec 09 18:18:34 crc kubenswrapper[4954]: I1209 18:18:34.025540 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7"} err="failed to get container status \"16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7\": rpc error: code = NotFound desc = could not find container \"16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7\": container with ID starting with 16a41d3ce090a51eeed0ee35be120a3e4ca6064b4b3088d76c9c8dbb30dbbab7 not found: ID does not exist" Dec 09 18:18:34 crc kubenswrapper[4954]: I1209 18:18:34.025561 4954 scope.go:117] "RemoveContainer" containerID="3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6" Dec 09 18:18:34 crc kubenswrapper[4954]: E1209 18:18:34.027230 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6\": container with ID starting with 3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6 not found: ID does not exist" containerID="3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6" Dec 09 18:18:34 crc kubenswrapper[4954]: I1209 18:18:34.027262 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6"} err="failed to get container status \"3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6\": rpc error: code = NotFound desc = could not find container \"3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6\": container with ID starting with 3ddd0994f762292a2e619aaaf35d6fda793bd0653db2d27ce309fe0ab3998be6 not found: ID does not exist" Dec 09 18:18:34 crc kubenswrapper[4954]: I1209 18:18:34.133570 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3efb54d9-560f-415c-b533-2236952dcca4" path="/var/lib/kubelet/pods/3efb54d9-560f-415c-b533-2236952dcca4/volumes" Dec 09 18:18:36 crc kubenswrapper[4954]: E1209 18:18:36.241427 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:18:36 crc kubenswrapper[4954]: E1209 18:18:36.241797 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:18:36 crc kubenswrapper[4954]: E1209 18:18:36.241935 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:18:36 crc kubenswrapper[4954]: E1209 18:18:36.243202 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:18:44 crc kubenswrapper[4954]: E1209 18:18:44.131004 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:18:48 crc kubenswrapper[4954]: E1209 18:18:48.124840 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:18:57 crc kubenswrapper[4954]: E1209 18:18:57.122513 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:19:01 crc kubenswrapper[4954]: E1209 18:19:01.124094 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:19:09 crc kubenswrapper[4954]: E1209 18:19:09.122338 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:19:15 crc kubenswrapper[4954]: E1209 18:19:15.122886 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:19:24 crc kubenswrapper[4954]: E1209 18:19:24.135097 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:19:26 crc kubenswrapper[4954]: E1209 18:19:26.123867 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:19:37 crc kubenswrapper[4954]: E1209 18:19:37.122209 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:19:39 crc kubenswrapper[4954]: E1209 18:19:39.123021 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:19:50 crc kubenswrapper[4954]: E1209 18:19:50.123988 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:19:51 crc kubenswrapper[4954]: E1209 18:19:51.122364 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:20:01 crc kubenswrapper[4954]: E1209 18:20:01.122886 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:20:04 crc kubenswrapper[4954]: E1209 18:20:04.129711 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:20:13 crc kubenswrapper[4954]: I1209 18:20:13.755088 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:20:13 crc kubenswrapper[4954]: I1209 18:20:13.755645 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:20:15 crc kubenswrapper[4954]: E1209 18:20:15.122507 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:20:15 crc kubenswrapper[4954]: E1209 18:20:15.122884 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:20:30 crc kubenswrapper[4954]: E1209 18:20:30.122540 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:20:30 crc kubenswrapper[4954]: E1209 18:20:30.122559 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:20:41 crc kubenswrapper[4954]: E1209 18:20:41.124258 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:20:43 crc kubenswrapper[4954]: I1209 18:20:43.755351 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:20:43 crc kubenswrapper[4954]: I1209 18:20:43.755747 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:20:44 crc kubenswrapper[4954]: E1209 18:20:44.132875 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:20:54 crc kubenswrapper[4954]: E1209 18:20:54.130492 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:20:57 crc kubenswrapper[4954]: E1209 18:20:57.122546 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:21:09 crc kubenswrapper[4954]: E1209 18:21:09.123473 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:21:12 crc kubenswrapper[4954]: E1209 18:21:12.122323 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:21:13 crc kubenswrapper[4954]: I1209 18:21:13.755443 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:21:13 crc kubenswrapper[4954]: I1209 18:21:13.755752 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:21:13 crc kubenswrapper[4954]: I1209 18:21:13.755799 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 18:21:13 crc kubenswrapper[4954]: I1209 18:21:13.756634 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f20ebcb5ad0c52185fb6ec22cec610c627bdbfc1f3c3a1e727e08a6d7a900c5b"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 18:21:13 crc kubenswrapper[4954]: I1209 18:21:13.756734 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://f20ebcb5ad0c52185fb6ec22cec610c627bdbfc1f3c3a1e727e08a6d7a900c5b" gracePeriod=600 Dec 09 18:21:14 crc kubenswrapper[4954]: I1209 18:21:14.446673 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="f20ebcb5ad0c52185fb6ec22cec610c627bdbfc1f3c3a1e727e08a6d7a900c5b" exitCode=0 Dec 09 18:21:14 crc kubenswrapper[4954]: I1209 18:21:14.446794 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"f20ebcb5ad0c52185fb6ec22cec610c627bdbfc1f3c3a1e727e08a6d7a900c5b"} Dec 09 18:21:14 crc kubenswrapper[4954]: I1209 18:21:14.447223 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353"} Dec 09 18:21:14 crc kubenswrapper[4954]: I1209 18:21:14.447242 4954 scope.go:117] "RemoveContainer" containerID="97defe02792ce3958d07190962edc307d96a1c4fff28563344dadfd006c7423e" Dec 09 18:21:20 crc kubenswrapper[4954]: E1209 18:21:20.122692 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:21:25 crc kubenswrapper[4954]: E1209 18:21:25.125450 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:21:32 crc kubenswrapper[4954]: E1209 18:21:32.122633 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:21:35 crc kubenswrapper[4954]: I1209 18:21:35.825904 4954 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="197b01d4-b9ac-4dc6-ac33-4638ece0241f" containerName="galera" probeResult="failure" output="command timed out" Dec 09 18:21:40 crc kubenswrapper[4954]: E1209 18:21:40.122878 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:21:44 crc kubenswrapper[4954]: E1209 18:21:44.122234 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:21:54 crc kubenswrapper[4954]: E1209 18:21:54.129588 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.048027 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bstch"] Dec 09 18:21:57 crc kubenswrapper[4954]: E1209 18:21:57.048996 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efb54d9-560f-415c-b533-2236952dcca4" containerName="registry-server" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.049016 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efb54d9-560f-415c-b533-2236952dcca4" containerName="registry-server" Dec 09 18:21:57 crc kubenswrapper[4954]: E1209 18:21:57.049039 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efb54d9-560f-415c-b533-2236952dcca4" containerName="extract-utilities" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.049047 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efb54d9-560f-415c-b533-2236952dcca4" containerName="extract-utilities" Dec 09 18:21:57 crc kubenswrapper[4954]: E1209 18:21:57.049075 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efb54d9-560f-415c-b533-2236952dcca4" containerName="extract-content" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.049083 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efb54d9-560f-415c-b533-2236952dcca4" containerName="extract-content" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.049402 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="3efb54d9-560f-415c-b533-2236952dcca4" containerName="registry-server" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.051555 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.058923 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bstch"] Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.217883 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-utilities\") pod \"community-operators-bstch\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.218042 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-catalog-content\") pod \"community-operators-bstch\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.218301 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn5ks\" (UniqueName: \"kubernetes.io/projected/d3065617-769b-4145-b155-7d35ee61c7f6-kube-api-access-mn5ks\") pod \"community-operators-bstch\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.320414 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn5ks\" (UniqueName: \"kubernetes.io/projected/d3065617-769b-4145-b155-7d35ee61c7f6-kube-api-access-mn5ks\") pod \"community-operators-bstch\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.320845 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-utilities\") pod \"community-operators-bstch\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.321090 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-catalog-content\") pod \"community-operators-bstch\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.321716 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-utilities\") pod \"community-operators-bstch\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.321774 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-catalog-content\") pod \"community-operators-bstch\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.352726 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn5ks\" (UniqueName: \"kubernetes.io/projected/d3065617-769b-4145-b155-7d35ee61c7f6-kube-api-access-mn5ks\") pod \"community-operators-bstch\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.380503 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bstch" Dec 09 18:21:57 crc kubenswrapper[4954]: I1209 18:21:57.941209 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bstch"] Dec 09 18:21:57 crc kubenswrapper[4954]: W1209 18:21:57.942410 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3065617_769b_4145_b155_7d35ee61c7f6.slice/crio-c56bc912b9e1bd26cad318d2b2a34864478a36c1a90bb9ac1d19f149a7e0c8ac WatchSource:0}: Error finding container c56bc912b9e1bd26cad318d2b2a34864478a36c1a90bb9ac1d19f149a7e0c8ac: Status 404 returned error can't find the container with id c56bc912b9e1bd26cad318d2b2a34864478a36c1a90bb9ac1d19f149a7e0c8ac Dec 09 18:21:58 crc kubenswrapper[4954]: I1209 18:21:58.872648 4954 generic.go:334] "Generic (PLEG): container finished" podID="d3065617-769b-4145-b155-7d35ee61c7f6" containerID="42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158" exitCode=0 Dec 09 18:21:58 crc kubenswrapper[4954]: I1209 18:21:58.872744 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstch" event={"ID":"d3065617-769b-4145-b155-7d35ee61c7f6","Type":"ContainerDied","Data":"42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158"} Dec 09 18:21:58 crc kubenswrapper[4954]: I1209 18:21:58.872943 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstch" event={"ID":"d3065617-769b-4145-b155-7d35ee61c7f6","Type":"ContainerStarted","Data":"c56bc912b9e1bd26cad318d2b2a34864478a36c1a90bb9ac1d19f149a7e0c8ac"} Dec 09 18:21:59 crc kubenswrapper[4954]: E1209 18:21:59.121768 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:22:01 crc kubenswrapper[4954]: I1209 18:22:01.533221 4954 generic.go:334] "Generic (PLEG): container finished" podID="d3065617-769b-4145-b155-7d35ee61c7f6" containerID="894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde" exitCode=0 Dec 09 18:22:01 crc kubenswrapper[4954]: I1209 18:22:01.533326 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstch" event={"ID":"d3065617-769b-4145-b155-7d35ee61c7f6","Type":"ContainerDied","Data":"894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde"} Dec 09 18:22:02 crc kubenswrapper[4954]: I1209 18:22:02.550383 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstch" event={"ID":"d3065617-769b-4145-b155-7d35ee61c7f6","Type":"ContainerStarted","Data":"43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1"} Dec 09 18:22:02 crc kubenswrapper[4954]: I1209 18:22:02.589362 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bstch" podStartSLOduration=2.485000225 podStartE2EDuration="5.589343727s" podCreationTimestamp="2025-12-09 18:21:57 +0000 UTC" firstStartedPulling="2025-12-09 18:21:58.879678859 +0000 UTC m=+5115.267852689" lastFinishedPulling="2025-12-09 18:22:01.984022371 +0000 UTC m=+5118.372196191" observedRunningTime="2025-12-09 18:22:02.580582243 +0000 UTC m=+5118.968756073" watchObservedRunningTime="2025-12-09 18:22:02.589343727 +0000 UTC m=+5118.977517547" Dec 09 18:22:06 crc kubenswrapper[4954]: E1209 18:22:06.122338 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:22:07 crc kubenswrapper[4954]: I1209 18:22:07.381029 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bstch" Dec 09 18:22:07 crc kubenswrapper[4954]: I1209 18:22:07.381353 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bstch" Dec 09 18:22:07 crc kubenswrapper[4954]: I1209 18:22:07.434118 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bstch" Dec 09 18:22:07 crc kubenswrapper[4954]: I1209 18:22:07.644919 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bstch" Dec 09 18:22:07 crc kubenswrapper[4954]: I1209 18:22:07.738954 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bstch"] Dec 09 18:22:09 crc kubenswrapper[4954]: I1209 18:22:09.615670 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bstch" podUID="d3065617-769b-4145-b155-7d35ee61c7f6" containerName="registry-server" containerID="cri-o://43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1" gracePeriod=2 Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.106889 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bstch" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.213687 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-utilities\") pod \"d3065617-769b-4145-b155-7d35ee61c7f6\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.213737 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-catalog-content\") pod \"d3065617-769b-4145-b155-7d35ee61c7f6\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.214012 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn5ks\" (UniqueName: \"kubernetes.io/projected/d3065617-769b-4145-b155-7d35ee61c7f6-kube-api-access-mn5ks\") pod \"d3065617-769b-4145-b155-7d35ee61c7f6\" (UID: \"d3065617-769b-4145-b155-7d35ee61c7f6\") " Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.214815 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-utilities" (OuterVolumeSpecName: "utilities") pod "d3065617-769b-4145-b155-7d35ee61c7f6" (UID: "d3065617-769b-4145-b155-7d35ee61c7f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.215474 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.220208 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3065617-769b-4145-b155-7d35ee61c7f6-kube-api-access-mn5ks" (OuterVolumeSpecName: "kube-api-access-mn5ks") pod "d3065617-769b-4145-b155-7d35ee61c7f6" (UID: "d3065617-769b-4145-b155-7d35ee61c7f6"). InnerVolumeSpecName "kube-api-access-mn5ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.266463 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3065617-769b-4145-b155-7d35ee61c7f6" (UID: "d3065617-769b-4145-b155-7d35ee61c7f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.317970 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn5ks\" (UniqueName: \"kubernetes.io/projected/d3065617-769b-4145-b155-7d35ee61c7f6-kube-api-access-mn5ks\") on node \"crc\" DevicePath \"\"" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.318012 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3065617-769b-4145-b155-7d35ee61c7f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.627202 4954 generic.go:334] "Generic (PLEG): container finished" podID="d3065617-769b-4145-b155-7d35ee61c7f6" containerID="43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1" exitCode=0 Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.627308 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstch" event={"ID":"d3065617-769b-4145-b155-7d35ee61c7f6","Type":"ContainerDied","Data":"43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1"} Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.627575 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bstch" event={"ID":"d3065617-769b-4145-b155-7d35ee61c7f6","Type":"ContainerDied","Data":"c56bc912b9e1bd26cad318d2b2a34864478a36c1a90bb9ac1d19f149a7e0c8ac"} Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.627610 4954 scope.go:117] "RemoveContainer" containerID="43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.627396 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bstch" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.647725 4954 scope.go:117] "RemoveContainer" containerID="894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.663736 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bstch"] Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.673276 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bstch"] Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.691322 4954 scope.go:117] "RemoveContainer" containerID="42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.721839 4954 scope.go:117] "RemoveContainer" containerID="43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1" Dec 09 18:22:10 crc kubenswrapper[4954]: E1209 18:22:10.722504 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1\": container with ID starting with 43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1 not found: ID does not exist" containerID="43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.722558 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1"} err="failed to get container status \"43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1\": rpc error: code = NotFound desc = could not find container \"43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1\": container with ID starting with 43509040460be9c4619a1811e10fe3b5e36f0cf42af335fb53f7d47874d333c1 not found: ID does not exist" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.722588 4954 scope.go:117] "RemoveContainer" containerID="894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde" Dec 09 18:22:10 crc kubenswrapper[4954]: E1209 18:22:10.723279 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde\": container with ID starting with 894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde not found: ID does not exist" containerID="894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.723324 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde"} err="failed to get container status \"894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde\": rpc error: code = NotFound desc = could not find container \"894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde\": container with ID starting with 894ffc6e4e84702e514831a748cf52b3aa6b7a391ee120e0bac71216c0de6dde not found: ID does not exist" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.723377 4954 scope.go:117] "RemoveContainer" containerID="42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158" Dec 09 18:22:10 crc kubenswrapper[4954]: E1209 18:22:10.723696 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158\": container with ID starting with 42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158 not found: ID does not exist" containerID="42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158" Dec 09 18:22:10 crc kubenswrapper[4954]: I1209 18:22:10.723731 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158"} err="failed to get container status \"42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158\": rpc error: code = NotFound desc = could not find container \"42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158\": container with ID starting with 42f5a923a7e8cd49a8adf0072cbbacafea6bba31da6d6b6ee611923463589158 not found: ID does not exist" Dec 09 18:22:12 crc kubenswrapper[4954]: I1209 18:22:12.130898 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3065617-769b-4145-b155-7d35ee61c7f6" path="/var/lib/kubelet/pods/d3065617-769b-4145-b155-7d35ee61c7f6/volumes" Dec 09 18:22:13 crc kubenswrapper[4954]: E1209 18:22:13.123166 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:22:20 crc kubenswrapper[4954]: E1209 18:22:20.121485 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:22:24 crc kubenswrapper[4954]: E1209 18:22:24.132367 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.078884 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dn72q"] Dec 09 18:22:32 crc kubenswrapper[4954]: E1209 18:22:32.080315 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3065617-769b-4145-b155-7d35ee61c7f6" containerName="extract-content" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.080339 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3065617-769b-4145-b155-7d35ee61c7f6" containerName="extract-content" Dec 09 18:22:32 crc kubenswrapper[4954]: E1209 18:22:32.080361 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3065617-769b-4145-b155-7d35ee61c7f6" containerName="registry-server" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.080369 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3065617-769b-4145-b155-7d35ee61c7f6" containerName="registry-server" Dec 09 18:22:32 crc kubenswrapper[4954]: E1209 18:22:32.080408 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3065617-769b-4145-b155-7d35ee61c7f6" containerName="extract-utilities" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.080417 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3065617-769b-4145-b155-7d35ee61c7f6" containerName="extract-utilities" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.080722 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3065617-769b-4145-b155-7d35ee61c7f6" containerName="registry-server" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.084475 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.108651 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dn72q"] Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.230162 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-catalog-content\") pod \"certified-operators-dn72q\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.230297 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-utilities\") pod \"certified-operators-dn72q\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.230434 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88kdr\" (UniqueName: \"kubernetes.io/projected/535a3ab4-a508-4ae9-ac16-94ac1e96a133-kube-api-access-88kdr\") pod \"certified-operators-dn72q\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.332184 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88kdr\" (UniqueName: \"kubernetes.io/projected/535a3ab4-a508-4ae9-ac16-94ac1e96a133-kube-api-access-88kdr\") pod \"certified-operators-dn72q\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.332338 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-catalog-content\") pod \"certified-operators-dn72q\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.332402 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-utilities\") pod \"certified-operators-dn72q\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.333043 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-utilities\") pod \"certified-operators-dn72q\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.333141 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-catalog-content\") pod \"certified-operators-dn72q\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.351725 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88kdr\" (UniqueName: \"kubernetes.io/projected/535a3ab4-a508-4ae9-ac16-94ac1e96a133-kube-api-access-88kdr\") pod \"certified-operators-dn72q\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:32 crc kubenswrapper[4954]: I1209 18:22:32.415739 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:33 crc kubenswrapper[4954]: I1209 18:22:33.026939 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dn72q"] Dec 09 18:22:33 crc kubenswrapper[4954]: E1209 18:22:33.121110 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:22:33 crc kubenswrapper[4954]: I1209 18:22:33.858351 4954 generic.go:334] "Generic (PLEG): container finished" podID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerID="0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad" exitCode=0 Dec 09 18:22:33 crc kubenswrapper[4954]: I1209 18:22:33.858493 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn72q" event={"ID":"535a3ab4-a508-4ae9-ac16-94ac1e96a133","Type":"ContainerDied","Data":"0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad"} Dec 09 18:22:33 crc kubenswrapper[4954]: I1209 18:22:33.858713 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn72q" event={"ID":"535a3ab4-a508-4ae9-ac16-94ac1e96a133","Type":"ContainerStarted","Data":"2fc1ec63026ec5de739bebc3714911f03584eb48fc01c88381eb801034da3f33"} Dec 09 18:22:35 crc kubenswrapper[4954]: I1209 18:22:35.880763 4954 generic.go:334] "Generic (PLEG): container finished" podID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerID="0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861" exitCode=0 Dec 09 18:22:35 crc kubenswrapper[4954]: I1209 18:22:35.880832 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn72q" event={"ID":"535a3ab4-a508-4ae9-ac16-94ac1e96a133","Type":"ContainerDied","Data":"0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861"} Dec 09 18:22:37 crc kubenswrapper[4954]: I1209 18:22:37.936233 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn72q" event={"ID":"535a3ab4-a508-4ae9-ac16-94ac1e96a133","Type":"ContainerStarted","Data":"d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55"} Dec 09 18:22:37 crc kubenswrapper[4954]: I1209 18:22:37.970235 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dn72q" podStartSLOduration=2.996648128 podStartE2EDuration="5.970209806s" podCreationTimestamp="2025-12-09 18:22:32 +0000 UTC" firstStartedPulling="2025-12-09 18:22:33.86035446 +0000 UTC m=+5150.248528280" lastFinishedPulling="2025-12-09 18:22:36.833916138 +0000 UTC m=+5153.222089958" observedRunningTime="2025-12-09 18:22:37.955908346 +0000 UTC m=+5154.344082206" watchObservedRunningTime="2025-12-09 18:22:37.970209806 +0000 UTC m=+5154.358383636" Dec 09 18:22:38 crc kubenswrapper[4954]: E1209 18:22:38.124119 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:22:42 crc kubenswrapper[4954]: I1209 18:22:42.416122 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:42 crc kubenswrapper[4954]: I1209 18:22:42.416767 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:42 crc kubenswrapper[4954]: I1209 18:22:42.461653 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:43 crc kubenswrapper[4954]: I1209 18:22:43.028193 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:43 crc kubenswrapper[4954]: I1209 18:22:43.080334 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dn72q"] Dec 09 18:22:44 crc kubenswrapper[4954]: I1209 18:22:44.997996 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dn72q" podUID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerName="registry-server" containerID="cri-o://d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55" gracePeriod=2 Dec 09 18:22:45 crc kubenswrapper[4954]: I1209 18:22:45.542836 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:45 crc kubenswrapper[4954]: I1209 18:22:45.613744 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-catalog-content\") pod \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " Dec 09 18:22:45 crc kubenswrapper[4954]: I1209 18:22:45.613902 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88kdr\" (UniqueName: \"kubernetes.io/projected/535a3ab4-a508-4ae9-ac16-94ac1e96a133-kube-api-access-88kdr\") pod \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " Dec 09 18:22:45 crc kubenswrapper[4954]: I1209 18:22:45.614232 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-utilities\") pod \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\" (UID: \"535a3ab4-a508-4ae9-ac16-94ac1e96a133\") " Dec 09 18:22:45 crc kubenswrapper[4954]: I1209 18:22:45.615045 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-utilities" (OuterVolumeSpecName: "utilities") pod "535a3ab4-a508-4ae9-ac16-94ac1e96a133" (UID: "535a3ab4-a508-4ae9-ac16-94ac1e96a133"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:22:45 crc kubenswrapper[4954]: I1209 18:22:45.619805 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/535a3ab4-a508-4ae9-ac16-94ac1e96a133-kube-api-access-88kdr" (OuterVolumeSpecName: "kube-api-access-88kdr") pod "535a3ab4-a508-4ae9-ac16-94ac1e96a133" (UID: "535a3ab4-a508-4ae9-ac16-94ac1e96a133"). InnerVolumeSpecName "kube-api-access-88kdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:22:45 crc kubenswrapper[4954]: I1209 18:22:45.666140 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "535a3ab4-a508-4ae9-ac16-94ac1e96a133" (UID: "535a3ab4-a508-4ae9-ac16-94ac1e96a133"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:22:45 crc kubenswrapper[4954]: I1209 18:22:45.717406 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88kdr\" (UniqueName: \"kubernetes.io/projected/535a3ab4-a508-4ae9-ac16-94ac1e96a133-kube-api-access-88kdr\") on node \"crc\" DevicePath \"\"" Dec 09 18:22:45 crc kubenswrapper[4954]: I1209 18:22:45.717451 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:22:45 crc kubenswrapper[4954]: I1209 18:22:45.717461 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/535a3ab4-a508-4ae9-ac16-94ac1e96a133-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.011152 4954 generic.go:334] "Generic (PLEG): container finished" podID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerID="d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55" exitCode=0 Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.011203 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn72q" event={"ID":"535a3ab4-a508-4ae9-ac16-94ac1e96a133","Type":"ContainerDied","Data":"d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55"} Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.011219 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn72q" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.011245 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn72q" event={"ID":"535a3ab4-a508-4ae9-ac16-94ac1e96a133","Type":"ContainerDied","Data":"2fc1ec63026ec5de739bebc3714911f03584eb48fc01c88381eb801034da3f33"} Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.011268 4954 scope.go:117] "RemoveContainer" containerID="d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.038336 4954 scope.go:117] "RemoveContainer" containerID="0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.060875 4954 scope.go:117] "RemoveContainer" containerID="0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.070958 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dn72q"] Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.084337 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dn72q"] Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.118232 4954 scope.go:117] "RemoveContainer" containerID="d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55" Dec 09 18:22:46 crc kubenswrapper[4954]: E1209 18:22:46.118698 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55\": container with ID starting with d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55 not found: ID does not exist" containerID="d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.118737 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55"} err="failed to get container status \"d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55\": rpc error: code = NotFound desc = could not find container \"d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55\": container with ID starting with d5f2a287281987136ac98173c2b18bcc7af4d985e070ef1c70c3ddaa24b18d55 not found: ID does not exist" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.118762 4954 scope.go:117] "RemoveContainer" containerID="0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861" Dec 09 18:22:46 crc kubenswrapper[4954]: E1209 18:22:46.119138 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861\": container with ID starting with 0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861 not found: ID does not exist" containerID="0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.119160 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861"} err="failed to get container status \"0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861\": rpc error: code = NotFound desc = could not find container \"0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861\": container with ID starting with 0160ae7c08643e5b307811e4d8f797ae3f5785a43dfa3066e40a460134deb861 not found: ID does not exist" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.119178 4954 scope.go:117] "RemoveContainer" containerID="0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad" Dec 09 18:22:46 crc kubenswrapper[4954]: E1209 18:22:46.119407 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad\": container with ID starting with 0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad not found: ID does not exist" containerID="0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.119432 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad"} err="failed to get container status \"0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad\": rpc error: code = NotFound desc = could not find container \"0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad\": container with ID starting with 0133a4c815d22d4f44c3f44bf20ca7240144610012d598423b147512d455c9ad not found: ID does not exist" Dec 09 18:22:46 crc kubenswrapper[4954]: I1209 18:22:46.131389 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" path="/var/lib/kubelet/pods/535a3ab4-a508-4ae9-ac16-94ac1e96a133/volumes" Dec 09 18:22:47 crc kubenswrapper[4954]: E1209 18:22:47.121897 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:22:53 crc kubenswrapper[4954]: E1209 18:22:53.122495 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:23:02 crc kubenswrapper[4954]: E1209 18:23:02.122604 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:23:07 crc kubenswrapper[4954]: E1209 18:23:07.122162 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:23:15 crc kubenswrapper[4954]: E1209 18:23:15.122670 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:23:21 crc kubenswrapper[4954]: I1209 18:23:21.122662 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 18:23:21 crc kubenswrapper[4954]: E1209 18:23:21.225910 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:23:21 crc kubenswrapper[4954]: E1209 18:23:21.225965 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:23:21 crc kubenswrapper[4954]: E1209 18:23:21.226156 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:23:21 crc kubenswrapper[4954]: E1209 18:23:21.227399 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:23:28 crc kubenswrapper[4954]: E1209 18:23:28.124974 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:23:32 crc kubenswrapper[4954]: E1209 18:23:32.126053 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:23:41 crc kubenswrapper[4954]: E1209 18:23:41.235995 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:23:41 crc kubenswrapper[4954]: E1209 18:23:41.236618 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:23:41 crc kubenswrapper[4954]: E1209 18:23:41.236794 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:23:41 crc kubenswrapper[4954]: E1209 18:23:41.238036 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:23:43 crc kubenswrapper[4954]: E1209 18:23:43.124922 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:23:43 crc kubenswrapper[4954]: I1209 18:23:43.756117 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:23:43 crc kubenswrapper[4954]: I1209 18:23:43.756182 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:23:54 crc kubenswrapper[4954]: E1209 18:23:54.129540 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:23:56 crc kubenswrapper[4954]: E1209 18:23:56.122481 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:24:08 crc kubenswrapper[4954]: E1209 18:24:08.123059 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:24:11 crc kubenswrapper[4954]: E1209 18:24:11.122534 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:24:13 crc kubenswrapper[4954]: I1209 18:24:13.755693 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:24:13 crc kubenswrapper[4954]: I1209 18:24:13.756327 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:24:20 crc kubenswrapper[4954]: E1209 18:24:20.123747 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:24:22 crc kubenswrapper[4954]: E1209 18:24:22.121610 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:24:33 crc kubenswrapper[4954]: E1209 18:24:33.123298 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:24:34 crc kubenswrapper[4954]: E1209 18:24:34.132039 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:24:43 crc kubenswrapper[4954]: I1209 18:24:43.803677 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:24:43 crc kubenswrapper[4954]: I1209 18:24:43.804254 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:24:43 crc kubenswrapper[4954]: I1209 18:24:43.804297 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 18:24:43 crc kubenswrapper[4954]: I1209 18:24:43.805138 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 18:24:43 crc kubenswrapper[4954]: I1209 18:24:43.805206 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" gracePeriod=600 Dec 09 18:24:43 crc kubenswrapper[4954]: E1209 18:24:43.945647 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:24:44 crc kubenswrapper[4954]: I1209 18:24:44.136060 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" exitCode=0 Dec 09 18:24:44 crc kubenswrapper[4954]: I1209 18:24:44.136794 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353"} Dec 09 18:24:44 crc kubenswrapper[4954]: I1209 18:24:44.136924 4954 scope.go:117] "RemoveContainer" containerID="f20ebcb5ad0c52185fb6ec22cec610c627bdbfc1f3c3a1e727e08a6d7a900c5b" Dec 09 18:24:44 crc kubenswrapper[4954]: I1209 18:24:44.137401 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:24:44 crc kubenswrapper[4954]: E1209 18:24:44.137940 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:24:47 crc kubenswrapper[4954]: E1209 18:24:47.123241 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:24:49 crc kubenswrapper[4954]: E1209 18:24:49.122553 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:24:56 crc kubenswrapper[4954]: I1209 18:24:56.120834 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:24:56 crc kubenswrapper[4954]: E1209 18:24:56.121727 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:25:00 crc kubenswrapper[4954]: E1209 18:25:00.122457 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:25:00 crc kubenswrapper[4954]: E1209 18:25:00.122504 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:25:07 crc kubenswrapper[4954]: I1209 18:25:07.122884 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:25:07 crc kubenswrapper[4954]: E1209 18:25:07.124181 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:25:15 crc kubenswrapper[4954]: E1209 18:25:15.122838 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:25:15 crc kubenswrapper[4954]: E1209 18:25:15.122890 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:25:21 crc kubenswrapper[4954]: I1209 18:25:21.120135 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:25:21 crc kubenswrapper[4954]: E1209 18:25:21.120930 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:25:26 crc kubenswrapper[4954]: E1209 18:25:26.124428 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:25:28 crc kubenswrapper[4954]: E1209 18:25:28.121928 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:25:36 crc kubenswrapper[4954]: I1209 18:25:36.122075 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:25:36 crc kubenswrapper[4954]: E1209 18:25:36.123239 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:25:37 crc kubenswrapper[4954]: E1209 18:25:37.123115 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:25:43 crc kubenswrapper[4954]: E1209 18:25:43.122438 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:25:48 crc kubenswrapper[4954]: I1209 18:25:48.120389 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:25:48 crc kubenswrapper[4954]: E1209 18:25:48.121271 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:25:51 crc kubenswrapper[4954]: E1209 18:25:51.122547 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:25:54 crc kubenswrapper[4954]: E1209 18:25:54.129216 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:26:02 crc kubenswrapper[4954]: I1209 18:26:02.121327 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:26:02 crc kubenswrapper[4954]: E1209 18:26:02.122095 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:26:04 crc kubenswrapper[4954]: E1209 18:26:04.131675 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:26:05 crc kubenswrapper[4954]: E1209 18:26:05.122734 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:26:17 crc kubenswrapper[4954]: I1209 18:26:17.120388 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:26:17 crc kubenswrapper[4954]: E1209 18:26:17.121342 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:26:17 crc kubenswrapper[4954]: E1209 18:26:17.123125 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:26:18 crc kubenswrapper[4954]: E1209 18:26:18.122645 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.707748 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6wf94"] Dec 09 18:26:25 crc kubenswrapper[4954]: E1209 18:26:25.708873 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerName="extract-content" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.708892 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerName="extract-content" Dec 09 18:26:25 crc kubenswrapper[4954]: E1209 18:26:25.708919 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerName="extract-utilities" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.708927 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerName="extract-utilities" Dec 09 18:26:25 crc kubenswrapper[4954]: E1209 18:26:25.708955 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerName="registry-server" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.708966 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerName="registry-server" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.709260 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="535a3ab4-a508-4ae9-ac16-94ac1e96a133" containerName="registry-server" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.711532 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.723658 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6wf94"] Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.850896 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtvng\" (UniqueName: \"kubernetes.io/projected/5757ba46-ef91-4926-a7aa-0c7bdefc6990-kube-api-access-gtvng\") pod \"redhat-operators-6wf94\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.851239 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-utilities\") pod \"redhat-operators-6wf94\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.851293 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-catalog-content\") pod \"redhat-operators-6wf94\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.953189 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtvng\" (UniqueName: \"kubernetes.io/projected/5757ba46-ef91-4926-a7aa-0c7bdefc6990-kube-api-access-gtvng\") pod \"redhat-operators-6wf94\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.953279 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-utilities\") pod \"redhat-operators-6wf94\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.953326 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-catalog-content\") pod \"redhat-operators-6wf94\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.955219 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-utilities\") pod \"redhat-operators-6wf94\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.955493 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-catalog-content\") pod \"redhat-operators-6wf94\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:25 crc kubenswrapper[4954]: I1209 18:26:25.986035 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtvng\" (UniqueName: \"kubernetes.io/projected/5757ba46-ef91-4926-a7aa-0c7bdefc6990-kube-api-access-gtvng\") pod \"redhat-operators-6wf94\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:26 crc kubenswrapper[4954]: I1209 18:26:26.042643 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:26 crc kubenswrapper[4954]: I1209 18:26:26.585944 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6wf94"] Dec 09 18:26:26 crc kubenswrapper[4954]: W1209 18:26:26.589980 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5757ba46_ef91_4926_a7aa_0c7bdefc6990.slice/crio-bb5620ea4a8dc8a459f420f84cdb115a8a81d89bc0f8c2c7203f0cfb9079c56d WatchSource:0}: Error finding container bb5620ea4a8dc8a459f420f84cdb115a8a81d89bc0f8c2c7203f0cfb9079c56d: Status 404 returned error can't find the container with id bb5620ea4a8dc8a459f420f84cdb115a8a81d89bc0f8c2c7203f0cfb9079c56d Dec 09 18:26:27 crc kubenswrapper[4954]: I1209 18:26:27.164522 4954 generic.go:334] "Generic (PLEG): container finished" podID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerID="b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461" exitCode=0 Dec 09 18:26:27 crc kubenswrapper[4954]: I1209 18:26:27.164692 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wf94" event={"ID":"5757ba46-ef91-4926-a7aa-0c7bdefc6990","Type":"ContainerDied","Data":"b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461"} Dec 09 18:26:27 crc kubenswrapper[4954]: I1209 18:26:27.165019 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wf94" event={"ID":"5757ba46-ef91-4926-a7aa-0c7bdefc6990","Type":"ContainerStarted","Data":"bb5620ea4a8dc8a459f420f84cdb115a8a81d89bc0f8c2c7203f0cfb9079c56d"} Dec 09 18:26:28 crc kubenswrapper[4954]: I1209 18:26:28.176773 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wf94" event={"ID":"5757ba46-ef91-4926-a7aa-0c7bdefc6990","Type":"ContainerStarted","Data":"ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971"} Dec 09 18:26:31 crc kubenswrapper[4954]: I1209 18:26:31.203372 4954 generic.go:334] "Generic (PLEG): container finished" podID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerID="ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971" exitCode=0 Dec 09 18:26:31 crc kubenswrapper[4954]: I1209 18:26:31.203442 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wf94" event={"ID":"5757ba46-ef91-4926-a7aa-0c7bdefc6990","Type":"ContainerDied","Data":"ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971"} Dec 09 18:26:32 crc kubenswrapper[4954]: I1209 18:26:32.121713 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:26:32 crc kubenswrapper[4954]: E1209 18:26:32.122153 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:26:32 crc kubenswrapper[4954]: E1209 18:26:32.122265 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:26:32 crc kubenswrapper[4954]: E1209 18:26:32.122327 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:26:32 crc kubenswrapper[4954]: I1209 18:26:32.216490 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wf94" event={"ID":"5757ba46-ef91-4926-a7aa-0c7bdefc6990","Type":"ContainerStarted","Data":"22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54"} Dec 09 18:26:32 crc kubenswrapper[4954]: I1209 18:26:32.244971 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6wf94" podStartSLOduration=2.692279729 podStartE2EDuration="7.244951452s" podCreationTimestamp="2025-12-09 18:26:25 +0000 UTC" firstStartedPulling="2025-12-09 18:26:27.166543355 +0000 UTC m=+5383.554717175" lastFinishedPulling="2025-12-09 18:26:31.719215078 +0000 UTC m=+5388.107388898" observedRunningTime="2025-12-09 18:26:32.237090095 +0000 UTC m=+5388.625263915" watchObservedRunningTime="2025-12-09 18:26:32.244951452 +0000 UTC m=+5388.633125272" Dec 09 18:26:36 crc kubenswrapper[4954]: I1209 18:26:36.043494 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:36 crc kubenswrapper[4954]: I1209 18:26:36.044049 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:37 crc kubenswrapper[4954]: I1209 18:26:37.092183 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6wf94" podUID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerName="registry-server" probeResult="failure" output=< Dec 09 18:26:37 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 18:26:37 crc kubenswrapper[4954]: > Dec 09 18:26:43 crc kubenswrapper[4954]: I1209 18:26:43.120206 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:26:43 crc kubenswrapper[4954]: E1209 18:26:43.122938 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:26:46 crc kubenswrapper[4954]: E1209 18:26:46.123336 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:26:46 crc kubenswrapper[4954]: I1209 18:26:46.263125 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:46 crc kubenswrapper[4954]: I1209 18:26:46.311678 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:46 crc kubenswrapper[4954]: I1209 18:26:46.503478 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6wf94"] Dec 09 18:26:47 crc kubenswrapper[4954]: E1209 18:26:47.121787 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:26:47 crc kubenswrapper[4954]: I1209 18:26:47.381631 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6wf94" podUID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerName="registry-server" containerID="cri-o://22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54" gracePeriod=2 Dec 09 18:26:47 crc kubenswrapper[4954]: I1209 18:26:47.866604 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:47 crc kubenswrapper[4954]: I1209 18:26:47.954292 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-utilities\") pod \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " Dec 09 18:26:47 crc kubenswrapper[4954]: I1209 18:26:47.954357 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtvng\" (UniqueName: \"kubernetes.io/projected/5757ba46-ef91-4926-a7aa-0c7bdefc6990-kube-api-access-gtvng\") pod \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " Dec 09 18:26:47 crc kubenswrapper[4954]: I1209 18:26:47.954522 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-catalog-content\") pod \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\" (UID: \"5757ba46-ef91-4926-a7aa-0c7bdefc6990\") " Dec 09 18:26:47 crc kubenswrapper[4954]: I1209 18:26:47.955496 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-utilities" (OuterVolumeSpecName: "utilities") pod "5757ba46-ef91-4926-a7aa-0c7bdefc6990" (UID: "5757ba46-ef91-4926-a7aa-0c7bdefc6990"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:26:47 crc kubenswrapper[4954]: I1209 18:26:47.960081 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5757ba46-ef91-4926-a7aa-0c7bdefc6990-kube-api-access-gtvng" (OuterVolumeSpecName: "kube-api-access-gtvng") pod "5757ba46-ef91-4926-a7aa-0c7bdefc6990" (UID: "5757ba46-ef91-4926-a7aa-0c7bdefc6990"). InnerVolumeSpecName "kube-api-access-gtvng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.057938 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.057985 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtvng\" (UniqueName: \"kubernetes.io/projected/5757ba46-ef91-4926-a7aa-0c7bdefc6990-kube-api-access-gtvng\") on node \"crc\" DevicePath \"\"" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.061053 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5757ba46-ef91-4926-a7aa-0c7bdefc6990" (UID: "5757ba46-ef91-4926-a7aa-0c7bdefc6990"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.160409 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5757ba46-ef91-4926-a7aa-0c7bdefc6990-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.394706 4954 generic.go:334] "Generic (PLEG): container finished" podID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerID="22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54" exitCode=0 Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.394751 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wf94" event={"ID":"5757ba46-ef91-4926-a7aa-0c7bdefc6990","Type":"ContainerDied","Data":"22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54"} Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.394806 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wf94" event={"ID":"5757ba46-ef91-4926-a7aa-0c7bdefc6990","Type":"ContainerDied","Data":"bb5620ea4a8dc8a459f420f84cdb115a8a81d89bc0f8c2c7203f0cfb9079c56d"} Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.394818 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6wf94" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.394828 4954 scope.go:117] "RemoveContainer" containerID="22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.417880 4954 scope.go:117] "RemoveContainer" containerID="ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.439231 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6wf94"] Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.447813 4954 scope.go:117] "RemoveContainer" containerID="b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.470232 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6wf94"] Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.547438 4954 scope.go:117] "RemoveContainer" containerID="22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54" Dec 09 18:26:48 crc kubenswrapper[4954]: E1209 18:26:48.548181 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54\": container with ID starting with 22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54 not found: ID does not exist" containerID="22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.548226 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54"} err="failed to get container status \"22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54\": rpc error: code = NotFound desc = could not find container \"22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54\": container with ID starting with 22a5a31e32c7113d53c16db5aa8190905163216e7e6842a481fa98aa3511bd54 not found: ID does not exist" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.548250 4954 scope.go:117] "RemoveContainer" containerID="ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971" Dec 09 18:26:48 crc kubenswrapper[4954]: E1209 18:26:48.548554 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971\": container with ID starting with ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971 not found: ID does not exist" containerID="ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.548568 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971"} err="failed to get container status \"ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971\": rpc error: code = NotFound desc = could not find container \"ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971\": container with ID starting with ad461e3ab96735d853f102cb83eadaa065b50b5a506bdfe879874b25cfbbb971 not found: ID does not exist" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.548579 4954 scope.go:117] "RemoveContainer" containerID="b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461" Dec 09 18:26:48 crc kubenswrapper[4954]: E1209 18:26:48.548929 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461\": container with ID starting with b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461 not found: ID does not exist" containerID="b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461" Dec 09 18:26:48 crc kubenswrapper[4954]: I1209 18:26:48.548950 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461"} err="failed to get container status \"b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461\": rpc error: code = NotFound desc = could not find container \"b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461\": container with ID starting with b838200dba384fd5386a90377274ba226b024972a0db63ca547d1ba0d2f42461 not found: ID does not exist" Dec 09 18:26:50 crc kubenswrapper[4954]: I1209 18:26:50.140740 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" path="/var/lib/kubelet/pods/5757ba46-ef91-4926-a7aa-0c7bdefc6990/volumes" Dec 09 18:26:54 crc kubenswrapper[4954]: I1209 18:26:54.136073 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:26:54 crc kubenswrapper[4954]: E1209 18:26:54.138057 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:26:57 crc kubenswrapper[4954]: E1209 18:26:57.122040 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:27:00 crc kubenswrapper[4954]: E1209 18:27:00.124452 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:27:06 crc kubenswrapper[4954]: I1209 18:27:06.120077 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:27:06 crc kubenswrapper[4954]: E1209 18:27:06.120882 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:27:09 crc kubenswrapper[4954]: E1209 18:27:09.122542 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:27:12 crc kubenswrapper[4954]: E1209 18:27:12.123147 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:27:19 crc kubenswrapper[4954]: I1209 18:27:19.120479 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:27:19 crc kubenswrapper[4954]: E1209 18:27:19.121689 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:27:23 crc kubenswrapper[4954]: E1209 18:27:23.144118 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:27:23 crc kubenswrapper[4954]: E1209 18:27:23.144118 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:27:34 crc kubenswrapper[4954]: I1209 18:27:34.128264 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:27:34 crc kubenswrapper[4954]: E1209 18:27:34.129761 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:27:35 crc kubenswrapper[4954]: E1209 18:27:35.122825 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:27:37 crc kubenswrapper[4954]: E1209 18:27:37.121796 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:27:45 crc kubenswrapper[4954]: I1209 18:27:45.120668 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:27:45 crc kubenswrapper[4954]: E1209 18:27:45.121502 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:27:48 crc kubenswrapper[4954]: E1209 18:27:48.126192 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:27:50 crc kubenswrapper[4954]: E1209 18:27:50.122373 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:28:00 crc kubenswrapper[4954]: I1209 18:28:00.121838 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:28:00 crc kubenswrapper[4954]: E1209 18:28:00.122730 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:28:00 crc kubenswrapper[4954]: E1209 18:28:00.124735 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:28:01 crc kubenswrapper[4954]: E1209 18:28:01.122787 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:28:11 crc kubenswrapper[4954]: E1209 18:28:11.122899 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:28:15 crc kubenswrapper[4954]: I1209 18:28:15.120548 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:28:15 crc kubenswrapper[4954]: E1209 18:28:15.121487 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:28:16 crc kubenswrapper[4954]: E1209 18:28:16.123378 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:28:22 crc kubenswrapper[4954]: I1209 18:28:22.123512 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 18:28:22 crc kubenswrapper[4954]: E1209 18:28:22.221902 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:28:22 crc kubenswrapper[4954]: E1209 18:28:22.221961 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:28:22 crc kubenswrapper[4954]: E1209 18:28:22.222160 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:28:22 crc kubenswrapper[4954]: E1209 18:28:22.223405 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:28:29 crc kubenswrapper[4954]: I1209 18:28:29.120864 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:28:29 crc kubenswrapper[4954]: E1209 18:28:29.121777 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:28:29 crc kubenswrapper[4954]: E1209 18:28:29.123282 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:28:34 crc kubenswrapper[4954]: E1209 18:28:34.130450 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:28:40 crc kubenswrapper[4954]: I1209 18:28:40.120500 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:28:40 crc kubenswrapper[4954]: E1209 18:28:40.121319 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:28:43 crc kubenswrapper[4954]: E1209 18:28:43.208090 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:28:43 crc kubenswrapper[4954]: E1209 18:28:43.209394 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:28:43 crc kubenswrapper[4954]: E1209 18:28:43.209687 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:28:43 crc kubenswrapper[4954]: E1209 18:28:43.210986 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:28:49 crc kubenswrapper[4954]: E1209 18:28:49.122513 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:28:55 crc kubenswrapper[4954]: I1209 18:28:55.122233 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:28:55 crc kubenswrapper[4954]: E1209 18:28:55.123261 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:28:56 crc kubenswrapper[4954]: E1209 18:28:56.123080 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:29:00 crc kubenswrapper[4954]: E1209 18:29:00.123630 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:29:07 crc kubenswrapper[4954]: E1209 18:29:07.122347 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:29:09 crc kubenswrapper[4954]: I1209 18:29:09.119865 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:29:09 crc kubenswrapper[4954]: E1209 18:29:09.120462 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:29:14 crc kubenswrapper[4954]: E1209 18:29:14.128985 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:29:18 crc kubenswrapper[4954]: E1209 18:29:18.124017 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:29:23 crc kubenswrapper[4954]: I1209 18:29:23.120898 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:29:23 crc kubenswrapper[4954]: E1209 18:29:23.121713 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:29:27 crc kubenswrapper[4954]: E1209 18:29:27.123466 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:29:31 crc kubenswrapper[4954]: E1209 18:29:31.123318 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.151838 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b8btm"] Dec 09 18:29:33 crc kubenswrapper[4954]: E1209 18:29:33.152779 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerName="extract-content" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.152795 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerName="extract-content" Dec 09 18:29:33 crc kubenswrapper[4954]: E1209 18:29:33.152814 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerName="registry-server" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.152820 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerName="registry-server" Dec 09 18:29:33 crc kubenswrapper[4954]: E1209 18:29:33.152842 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerName="extract-utilities" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.152848 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerName="extract-utilities" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.153070 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="5757ba46-ef91-4926-a7aa-0c7bdefc6990" containerName="registry-server" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.154896 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.167129 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8btm"] Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.334938 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-utilities\") pod \"redhat-marketplace-b8btm\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.335133 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88fbt\" (UniqueName: \"kubernetes.io/projected/d01a929c-18d0-4034-9437-17d520733ebc-kube-api-access-88fbt\") pod \"redhat-marketplace-b8btm\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.335227 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-catalog-content\") pod \"redhat-marketplace-b8btm\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.437152 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-utilities\") pod \"redhat-marketplace-b8btm\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.437233 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88fbt\" (UniqueName: \"kubernetes.io/projected/d01a929c-18d0-4034-9437-17d520733ebc-kube-api-access-88fbt\") pod \"redhat-marketplace-b8btm\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.437259 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-catalog-content\") pod \"redhat-marketplace-b8btm\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.437754 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-utilities\") pod \"redhat-marketplace-b8btm\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.437857 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-catalog-content\") pod \"redhat-marketplace-b8btm\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.457545 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88fbt\" (UniqueName: \"kubernetes.io/projected/d01a929c-18d0-4034-9437-17d520733ebc-kube-api-access-88fbt\") pod \"redhat-marketplace-b8btm\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:33 crc kubenswrapper[4954]: I1209 18:29:33.484576 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:34 crc kubenswrapper[4954]: I1209 18:29:34.009231 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8btm"] Dec 09 18:29:34 crc kubenswrapper[4954]: I1209 18:29:34.195084 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8btm" event={"ID":"d01a929c-18d0-4034-9437-17d520733ebc","Type":"ContainerStarted","Data":"17897b38ff7592ee8b674538997f52d5b328f47277aa65f0fdcc3034ccff7ada"} Dec 09 18:29:35 crc kubenswrapper[4954]: I1209 18:29:35.121247 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:29:35 crc kubenswrapper[4954]: E1209 18:29:35.121897 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:29:35 crc kubenswrapper[4954]: I1209 18:29:35.205369 4954 generic.go:334] "Generic (PLEG): container finished" podID="d01a929c-18d0-4034-9437-17d520733ebc" containerID="751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99" exitCode=0 Dec 09 18:29:35 crc kubenswrapper[4954]: I1209 18:29:35.205426 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8btm" event={"ID":"d01a929c-18d0-4034-9437-17d520733ebc","Type":"ContainerDied","Data":"751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99"} Dec 09 18:29:36 crc kubenswrapper[4954]: I1209 18:29:36.223035 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8btm" event={"ID":"d01a929c-18d0-4034-9437-17d520733ebc","Type":"ContainerStarted","Data":"2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d"} Dec 09 18:29:37 crc kubenswrapper[4954]: I1209 18:29:37.235237 4954 generic.go:334] "Generic (PLEG): container finished" podID="d01a929c-18d0-4034-9437-17d520733ebc" containerID="2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d" exitCode=0 Dec 09 18:29:37 crc kubenswrapper[4954]: I1209 18:29:37.235281 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8btm" event={"ID":"d01a929c-18d0-4034-9437-17d520733ebc","Type":"ContainerDied","Data":"2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d"} Dec 09 18:29:38 crc kubenswrapper[4954]: I1209 18:29:38.245614 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8btm" event={"ID":"d01a929c-18d0-4034-9437-17d520733ebc","Type":"ContainerStarted","Data":"06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1"} Dec 09 18:29:38 crc kubenswrapper[4954]: I1209 18:29:38.271977 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b8btm" podStartSLOduration=2.583427812 podStartE2EDuration="5.271951155s" podCreationTimestamp="2025-12-09 18:29:33 +0000 UTC" firstStartedPulling="2025-12-09 18:29:35.207850717 +0000 UTC m=+5571.596024537" lastFinishedPulling="2025-12-09 18:29:37.89637406 +0000 UTC m=+5574.284547880" observedRunningTime="2025-12-09 18:29:38.262818287 +0000 UTC m=+5574.650992147" watchObservedRunningTime="2025-12-09 18:29:38.271951155 +0000 UTC m=+5574.660124995" Dec 09 18:29:41 crc kubenswrapper[4954]: E1209 18:29:41.122075 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:29:43 crc kubenswrapper[4954]: I1209 18:29:43.485688 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:43 crc kubenswrapper[4954]: I1209 18:29:43.486079 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:43 crc kubenswrapper[4954]: I1209 18:29:43.532283 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:44 crc kubenswrapper[4954]: I1209 18:29:44.351913 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:44 crc kubenswrapper[4954]: I1209 18:29:44.405758 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8btm"] Dec 09 18:29:45 crc kubenswrapper[4954]: E1209 18:29:45.122354 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:29:46 crc kubenswrapper[4954]: I1209 18:29:46.120888 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:29:46 crc kubenswrapper[4954]: I1209 18:29:46.313826 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b8btm" podUID="d01a929c-18d0-4034-9437-17d520733ebc" containerName="registry-server" containerID="cri-o://06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1" gracePeriod=2 Dec 09 18:29:46 crc kubenswrapper[4954]: I1209 18:29:46.870842 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:46 crc kubenswrapper[4954]: I1209 18:29:46.958226 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-catalog-content\") pod \"d01a929c-18d0-4034-9437-17d520733ebc\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " Dec 09 18:29:46 crc kubenswrapper[4954]: I1209 18:29:46.958514 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-utilities\") pod \"d01a929c-18d0-4034-9437-17d520733ebc\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " Dec 09 18:29:46 crc kubenswrapper[4954]: I1209 18:29:46.958578 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88fbt\" (UniqueName: \"kubernetes.io/projected/d01a929c-18d0-4034-9437-17d520733ebc-kube-api-access-88fbt\") pod \"d01a929c-18d0-4034-9437-17d520733ebc\" (UID: \"d01a929c-18d0-4034-9437-17d520733ebc\") " Dec 09 18:29:46 crc kubenswrapper[4954]: I1209 18:29:46.959725 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-utilities" (OuterVolumeSpecName: "utilities") pod "d01a929c-18d0-4034-9437-17d520733ebc" (UID: "d01a929c-18d0-4034-9437-17d520733ebc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:29:46 crc kubenswrapper[4954]: I1209 18:29:46.965814 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d01a929c-18d0-4034-9437-17d520733ebc-kube-api-access-88fbt" (OuterVolumeSpecName: "kube-api-access-88fbt") pod "d01a929c-18d0-4034-9437-17d520733ebc" (UID: "d01a929c-18d0-4034-9437-17d520733ebc"). InnerVolumeSpecName "kube-api-access-88fbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:29:46 crc kubenswrapper[4954]: I1209 18:29:46.985491 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d01a929c-18d0-4034-9437-17d520733ebc" (UID: "d01a929c-18d0-4034-9437-17d520733ebc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.059949 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.059976 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d01a929c-18d0-4034-9437-17d520733ebc-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.059986 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88fbt\" (UniqueName: \"kubernetes.io/projected/d01a929c-18d0-4034-9437-17d520733ebc-kube-api-access-88fbt\") on node \"crc\" DevicePath \"\"" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.324832 4954 generic.go:334] "Generic (PLEG): container finished" podID="d01a929c-18d0-4034-9437-17d520733ebc" containerID="06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1" exitCode=0 Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.325020 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8btm" event={"ID":"d01a929c-18d0-4034-9437-17d520733ebc","Type":"ContainerDied","Data":"06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1"} Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.325221 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8btm" event={"ID":"d01a929c-18d0-4034-9437-17d520733ebc","Type":"ContainerDied","Data":"17897b38ff7592ee8b674538997f52d5b328f47277aa65f0fdcc3034ccff7ada"} Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.325244 4954 scope.go:117] "RemoveContainer" containerID="06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.325123 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8btm" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.328358 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"afd4d72dcad9520c4cb369ae68cbac6019ac340837b03fa2d274871625d5a7a8"} Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.347888 4954 scope.go:117] "RemoveContainer" containerID="2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.381712 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8btm"] Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.382395 4954 scope.go:117] "RemoveContainer" containerID="751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.393788 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8btm"] Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.427561 4954 scope.go:117] "RemoveContainer" containerID="06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1" Dec 09 18:29:47 crc kubenswrapper[4954]: E1209 18:29:47.428008 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1\": container with ID starting with 06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1 not found: ID does not exist" containerID="06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.428045 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1"} err="failed to get container status \"06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1\": rpc error: code = NotFound desc = could not find container \"06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1\": container with ID starting with 06bff0c5ac01a27e7fe81f1e18c132032db2d92b3354c2670a5d8f21a40661d1 not found: ID does not exist" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.428072 4954 scope.go:117] "RemoveContainer" containerID="2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d" Dec 09 18:29:47 crc kubenswrapper[4954]: E1209 18:29:47.428428 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d\": container with ID starting with 2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d not found: ID does not exist" containerID="2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.428454 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d"} err="failed to get container status \"2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d\": rpc error: code = NotFound desc = could not find container \"2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d\": container with ID starting with 2f5e5c0f52731992cb8b443478eb0ecbce4f604fbe2cf926af93341a0291492d not found: ID does not exist" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.428475 4954 scope.go:117] "RemoveContainer" containerID="751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99" Dec 09 18:29:47 crc kubenswrapper[4954]: E1209 18:29:47.428898 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99\": container with ID starting with 751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99 not found: ID does not exist" containerID="751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99" Dec 09 18:29:47 crc kubenswrapper[4954]: I1209 18:29:47.428923 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99"} err="failed to get container status \"751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99\": rpc error: code = NotFound desc = could not find container \"751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99\": container with ID starting with 751c783d522ba202d41a48ee4a272a1f9ad6a6cefa7941c2c149f435ad8b0d99 not found: ID does not exist" Dec 09 18:29:48 crc kubenswrapper[4954]: I1209 18:29:48.142542 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d01a929c-18d0-4034-9437-17d520733ebc" path="/var/lib/kubelet/pods/d01a929c-18d0-4034-9437-17d520733ebc/volumes" Dec 09 18:29:55 crc kubenswrapper[4954]: E1209 18:29:55.122762 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:29:58 crc kubenswrapper[4954]: E1209 18:29:58.124509 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.153407 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr"] Dec 09 18:30:00 crc kubenswrapper[4954]: E1209 18:30:00.154355 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d01a929c-18d0-4034-9437-17d520733ebc" containerName="extract-utilities" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.154371 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d01a929c-18d0-4034-9437-17d520733ebc" containerName="extract-utilities" Dec 09 18:30:00 crc kubenswrapper[4954]: E1209 18:30:00.154389 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d01a929c-18d0-4034-9437-17d520733ebc" containerName="registry-server" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.154395 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d01a929c-18d0-4034-9437-17d520733ebc" containerName="registry-server" Dec 09 18:30:00 crc kubenswrapper[4954]: E1209 18:30:00.154423 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d01a929c-18d0-4034-9437-17d520733ebc" containerName="extract-content" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.154431 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d01a929c-18d0-4034-9437-17d520733ebc" containerName="extract-content" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.154717 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d01a929c-18d0-4034-9437-17d520733ebc" containerName="registry-server" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.155645 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.158354 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.158729 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.166226 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr"] Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.336509 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwb9p\" (UniqueName: \"kubernetes.io/projected/bbfb441d-2f1b-4116-8960-7d7b46d7300c-kube-api-access-qwb9p\") pod \"collect-profiles-29421750-nlqbr\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.337171 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbfb441d-2f1b-4116-8960-7d7b46d7300c-secret-volume\") pod \"collect-profiles-29421750-nlqbr\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.337230 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbfb441d-2f1b-4116-8960-7d7b46d7300c-config-volume\") pod \"collect-profiles-29421750-nlqbr\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.438849 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbfb441d-2f1b-4116-8960-7d7b46d7300c-secret-volume\") pod \"collect-profiles-29421750-nlqbr\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.438934 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbfb441d-2f1b-4116-8960-7d7b46d7300c-config-volume\") pod \"collect-profiles-29421750-nlqbr\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.439116 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwb9p\" (UniqueName: \"kubernetes.io/projected/bbfb441d-2f1b-4116-8960-7d7b46d7300c-kube-api-access-qwb9p\") pod \"collect-profiles-29421750-nlqbr\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.440816 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbfb441d-2f1b-4116-8960-7d7b46d7300c-config-volume\") pod \"collect-profiles-29421750-nlqbr\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.447713 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbfb441d-2f1b-4116-8960-7d7b46d7300c-secret-volume\") pod \"collect-profiles-29421750-nlqbr\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.461644 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwb9p\" (UniqueName: \"kubernetes.io/projected/bbfb441d-2f1b-4116-8960-7d7b46d7300c-kube-api-access-qwb9p\") pod \"collect-profiles-29421750-nlqbr\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:00 crc kubenswrapper[4954]: I1209 18:30:00.484889 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:01 crc kubenswrapper[4954]: I1209 18:30:01.297693 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr"] Dec 09 18:30:01 crc kubenswrapper[4954]: I1209 18:30:01.474393 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" event={"ID":"bbfb441d-2f1b-4116-8960-7d7b46d7300c","Type":"ContainerStarted","Data":"f20f26830bc102bb5e23eb9d6b3f0a25fa58dea1970c35f95d16ab688413d13b"} Dec 09 18:30:02 crc kubenswrapper[4954]: I1209 18:30:02.490324 4954 generic.go:334] "Generic (PLEG): container finished" podID="bbfb441d-2f1b-4116-8960-7d7b46d7300c" containerID="fd3adcd03d989e9f1ef6c1cdbca7792aaad7e087a597d85762da11f330b21ab3" exitCode=0 Dec 09 18:30:02 crc kubenswrapper[4954]: I1209 18:30:02.490773 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" event={"ID":"bbfb441d-2f1b-4116-8960-7d7b46d7300c","Type":"ContainerDied","Data":"fd3adcd03d989e9f1ef6c1cdbca7792aaad7e087a597d85762da11f330b21ab3"} Dec 09 18:30:03 crc kubenswrapper[4954]: I1209 18:30:03.861489 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.019424 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbfb441d-2f1b-4116-8960-7d7b46d7300c-secret-volume\") pod \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.019526 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwb9p\" (UniqueName: \"kubernetes.io/projected/bbfb441d-2f1b-4116-8960-7d7b46d7300c-kube-api-access-qwb9p\") pod \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.019583 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbfb441d-2f1b-4116-8960-7d7b46d7300c-config-volume\") pod \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\" (UID: \"bbfb441d-2f1b-4116-8960-7d7b46d7300c\") " Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.020304 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbfb441d-2f1b-4116-8960-7d7b46d7300c-config-volume" (OuterVolumeSpecName: "config-volume") pod "bbfb441d-2f1b-4116-8960-7d7b46d7300c" (UID: "bbfb441d-2f1b-4116-8960-7d7b46d7300c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.021275 4954 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbfb441d-2f1b-4116-8960-7d7b46d7300c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.028841 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbfb441d-2f1b-4116-8960-7d7b46d7300c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bbfb441d-2f1b-4116-8960-7d7b46d7300c" (UID: "bbfb441d-2f1b-4116-8960-7d7b46d7300c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.029053 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbfb441d-2f1b-4116-8960-7d7b46d7300c-kube-api-access-qwb9p" (OuterVolumeSpecName: "kube-api-access-qwb9p") pod "bbfb441d-2f1b-4116-8960-7d7b46d7300c" (UID: "bbfb441d-2f1b-4116-8960-7d7b46d7300c"). InnerVolumeSpecName "kube-api-access-qwb9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.123017 4954 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbfb441d-2f1b-4116-8960-7d7b46d7300c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.123057 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwb9p\" (UniqueName: \"kubernetes.io/projected/bbfb441d-2f1b-4116-8960-7d7b46d7300c-kube-api-access-qwb9p\") on node \"crc\" DevicePath \"\"" Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.511335 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" event={"ID":"bbfb441d-2f1b-4116-8960-7d7b46d7300c","Type":"ContainerDied","Data":"f20f26830bc102bb5e23eb9d6b3f0a25fa58dea1970c35f95d16ab688413d13b"} Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.511376 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f20f26830bc102bb5e23eb9d6b3f0a25fa58dea1970c35f95d16ab688413d13b" Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.511380 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421750-nlqbr" Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.938332 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg"] Dec 09 18:30:04 crc kubenswrapper[4954]: I1209 18:30:04.951781 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421705-hj5wg"] Dec 09 18:30:06 crc kubenswrapper[4954]: I1209 18:30:06.133733 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6882048a-c4d8-4669-a3c0-262dad0a4fed" path="/var/lib/kubelet/pods/6882048a-c4d8-4669-a3c0-262dad0a4fed/volumes" Dec 09 18:30:06 crc kubenswrapper[4954]: I1209 18:30:06.805331 4954 scope.go:117] "RemoveContainer" containerID="9727c9686cd96a73887834781ed2a0ea0ddbe396363e339e075bd0f690b60c2c" Dec 09 18:30:08 crc kubenswrapper[4954]: E1209 18:30:08.122831 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:30:11 crc kubenswrapper[4954]: E1209 18:30:11.123482 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:30:23 crc kubenswrapper[4954]: E1209 18:30:23.123683 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:30:25 crc kubenswrapper[4954]: E1209 18:30:25.121942 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:30:38 crc kubenswrapper[4954]: E1209 18:30:38.123452 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:30:39 crc kubenswrapper[4954]: E1209 18:30:39.122010 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:30:53 crc kubenswrapper[4954]: E1209 18:30:53.124548 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:30:54 crc kubenswrapper[4954]: E1209 18:30:54.131432 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:31:05 crc kubenswrapper[4954]: E1209 18:31:05.123089 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:31:07 crc kubenswrapper[4954]: E1209 18:31:07.121866 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:31:19 crc kubenswrapper[4954]: E1209 18:31:19.123223 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:31:22 crc kubenswrapper[4954]: E1209 18:31:22.123457 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:31:30 crc kubenswrapper[4954]: E1209 18:31:30.122538 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:31:33 crc kubenswrapper[4954]: E1209 18:31:33.123132 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:31:44 crc kubenswrapper[4954]: E1209 18:31:44.133135 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:31:46 crc kubenswrapper[4954]: E1209 18:31:46.123485 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:31:57 crc kubenswrapper[4954]: E1209 18:31:57.123475 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:31:58 crc kubenswrapper[4954]: E1209 18:31:58.209078 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:32:09 crc kubenswrapper[4954]: E1209 18:32:09.122303 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:32:09 crc kubenswrapper[4954]: E1209 18:32:09.122302 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:32:13 crc kubenswrapper[4954]: I1209 18:32:13.755472 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:32:13 crc kubenswrapper[4954]: I1209 18:32:13.756091 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:32:20 crc kubenswrapper[4954]: E1209 18:32:20.126199 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:32:21 crc kubenswrapper[4954]: E1209 18:32:21.121626 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:32:31 crc kubenswrapper[4954]: E1209 18:32:31.122248 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:32:34 crc kubenswrapper[4954]: E1209 18:32:34.137150 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:32:43 crc kubenswrapper[4954]: E1209 18:32:43.123605 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:32:43 crc kubenswrapper[4954]: I1209 18:32:43.755986 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:32:43 crc kubenswrapper[4954]: I1209 18:32:43.756077 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.609999 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hlk28/must-gather-b4g8k"] Dec 09 18:32:44 crc kubenswrapper[4954]: E1209 18:32:44.611863 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbfb441d-2f1b-4116-8960-7d7b46d7300c" containerName="collect-profiles" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.611987 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbfb441d-2f1b-4116-8960-7d7b46d7300c" containerName="collect-profiles" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.612440 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbfb441d-2f1b-4116-8960-7d7b46d7300c" containerName="collect-profiles" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.614120 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/must-gather-b4g8k" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.616209 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hlk28"/"openshift-service-ca.crt" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.616574 4954 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hlk28"/"kube-root-ca.crt" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.642613 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hlk28/must-gather-b4g8k"] Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.661886 4954 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-hlk28"/"default-dockercfg-4fp5r" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.711093 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gc5c\" (UniqueName: \"kubernetes.io/projected/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-kube-api-access-4gc5c\") pod \"must-gather-b4g8k\" (UID: \"9b11b7cc-7b44-4800-95b1-37f4cd8a6146\") " pod="openshift-must-gather-hlk28/must-gather-b4g8k" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.712088 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-must-gather-output\") pod \"must-gather-b4g8k\" (UID: \"9b11b7cc-7b44-4800-95b1-37f4cd8a6146\") " pod="openshift-must-gather-hlk28/must-gather-b4g8k" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.813057 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gc5c\" (UniqueName: \"kubernetes.io/projected/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-kube-api-access-4gc5c\") pod \"must-gather-b4g8k\" (UID: \"9b11b7cc-7b44-4800-95b1-37f4cd8a6146\") " pod="openshift-must-gather-hlk28/must-gather-b4g8k" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.813481 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-must-gather-output\") pod \"must-gather-b4g8k\" (UID: \"9b11b7cc-7b44-4800-95b1-37f4cd8a6146\") " pod="openshift-must-gather-hlk28/must-gather-b4g8k" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.813926 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-must-gather-output\") pod \"must-gather-b4g8k\" (UID: \"9b11b7cc-7b44-4800-95b1-37f4cd8a6146\") " pod="openshift-must-gather-hlk28/must-gather-b4g8k" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.853725 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gc5c\" (UniqueName: \"kubernetes.io/projected/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-kube-api-access-4gc5c\") pod \"must-gather-b4g8k\" (UID: \"9b11b7cc-7b44-4800-95b1-37f4cd8a6146\") " pod="openshift-must-gather-hlk28/must-gather-b4g8k" Dec 09 18:32:44 crc kubenswrapper[4954]: I1209 18:32:44.946702 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/must-gather-b4g8k" Dec 09 18:32:45 crc kubenswrapper[4954]: I1209 18:32:45.535727 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hlk28/must-gather-b4g8k"] Dec 09 18:32:46 crc kubenswrapper[4954]: I1209 18:32:46.270245 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hlk28/must-gather-b4g8k" event={"ID":"9b11b7cc-7b44-4800-95b1-37f4cd8a6146","Type":"ContainerStarted","Data":"9df43768e673d15e90fb20b131dc41cb9ba1c59093dd504b7c0f579356379df4"} Dec 09 18:32:50 crc kubenswrapper[4954]: E1209 18:32:50.122372 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:32:53 crc kubenswrapper[4954]: I1209 18:32:53.349276 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hlk28/must-gather-b4g8k" event={"ID":"9b11b7cc-7b44-4800-95b1-37f4cd8a6146","Type":"ContainerStarted","Data":"3c165ce3af65b7ee339bdb0106e348eba8c01f6f953634bbaffcde8d99b22101"} Dec 09 18:32:53 crc kubenswrapper[4954]: I1209 18:32:53.349777 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hlk28/must-gather-b4g8k" event={"ID":"9b11b7cc-7b44-4800-95b1-37f4cd8a6146","Type":"ContainerStarted","Data":"146f170dd8360de9b5092b06ed52724ce53e6a215cfec261d8d8471a61271a20"} Dec 09 18:32:53 crc kubenswrapper[4954]: I1209 18:32:53.369455 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hlk28/must-gather-b4g8k" podStartSLOduration=2.518152517 podStartE2EDuration="9.369432381s" podCreationTimestamp="2025-12-09 18:32:44 +0000 UTC" firstStartedPulling="2025-12-09 18:32:45.546852147 +0000 UTC m=+5761.935025967" lastFinishedPulling="2025-12-09 18:32:52.398132011 +0000 UTC m=+5768.786305831" observedRunningTime="2025-12-09 18:32:53.363715062 +0000 UTC m=+5769.751888882" watchObservedRunningTime="2025-12-09 18:32:53.369432381 +0000 UTC m=+5769.757606241" Dec 09 18:32:54 crc kubenswrapper[4954]: E1209 18:32:54.129889 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.617116 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hlk28/crc-debug-ggzjh"] Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.619016 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/crc-debug-ggzjh" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.687150 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c789fbb9-645e-447d-aef2-b3d22c3931f3-host\") pod \"crc-debug-ggzjh\" (UID: \"c789fbb9-645e-447d-aef2-b3d22c3931f3\") " pod="openshift-must-gather-hlk28/crc-debug-ggzjh" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.687353 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txxdh\" (UniqueName: \"kubernetes.io/projected/c789fbb9-645e-447d-aef2-b3d22c3931f3-kube-api-access-txxdh\") pod \"crc-debug-ggzjh\" (UID: \"c789fbb9-645e-447d-aef2-b3d22c3931f3\") " pod="openshift-must-gather-hlk28/crc-debug-ggzjh" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.789024 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c789fbb9-645e-447d-aef2-b3d22c3931f3-host\") pod \"crc-debug-ggzjh\" (UID: \"c789fbb9-645e-447d-aef2-b3d22c3931f3\") " pod="openshift-must-gather-hlk28/crc-debug-ggzjh" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.789168 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txxdh\" (UniqueName: \"kubernetes.io/projected/c789fbb9-645e-447d-aef2-b3d22c3931f3-kube-api-access-txxdh\") pod \"crc-debug-ggzjh\" (UID: \"c789fbb9-645e-447d-aef2-b3d22c3931f3\") " pod="openshift-must-gather-hlk28/crc-debug-ggzjh" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.789174 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c789fbb9-645e-447d-aef2-b3d22c3931f3-host\") pod \"crc-debug-ggzjh\" (UID: \"c789fbb9-645e-447d-aef2-b3d22c3931f3\") " pod="openshift-must-gather-hlk28/crc-debug-ggzjh" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.810915 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txxdh\" (UniqueName: \"kubernetes.io/projected/c789fbb9-645e-447d-aef2-b3d22c3931f3-kube-api-access-txxdh\") pod \"crc-debug-ggzjh\" (UID: \"c789fbb9-645e-447d-aef2-b3d22c3931f3\") " pod="openshift-must-gather-hlk28/crc-debug-ggzjh" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.909474 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zk2lx"] Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.912455 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.932913 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zk2lx"] Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.940868 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/crc-debug-ggzjh" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.996009 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxbhm\" (UniqueName: \"kubernetes.io/projected/cb9b49d8-6fc1-439a-8a16-f79936024d71-kube-api-access-vxbhm\") pod \"community-operators-zk2lx\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.996120 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-utilities\") pod \"community-operators-zk2lx\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:56 crc kubenswrapper[4954]: I1209 18:32:56.996215 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-catalog-content\") pod \"community-operators-zk2lx\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:57 crc kubenswrapper[4954]: W1209 18:32:57.049742 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc789fbb9_645e_447d_aef2_b3d22c3931f3.slice/crio-a98a6016458c944372a49ee1c1f44b768e4fae1a15069f0c1eb02564978eb512 WatchSource:0}: Error finding container a98a6016458c944372a49ee1c1f44b768e4fae1a15069f0c1eb02564978eb512: Status 404 returned error can't find the container with id a98a6016458c944372a49ee1c1f44b768e4fae1a15069f0c1eb02564978eb512 Dec 09 18:32:57 crc kubenswrapper[4954]: I1209 18:32:57.098398 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxbhm\" (UniqueName: \"kubernetes.io/projected/cb9b49d8-6fc1-439a-8a16-f79936024d71-kube-api-access-vxbhm\") pod \"community-operators-zk2lx\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:57 crc kubenswrapper[4954]: I1209 18:32:57.098488 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-utilities\") pod \"community-operators-zk2lx\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:57 crc kubenswrapper[4954]: I1209 18:32:57.098553 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-catalog-content\") pod \"community-operators-zk2lx\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:57 crc kubenswrapper[4954]: I1209 18:32:57.099042 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-catalog-content\") pod \"community-operators-zk2lx\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:57 crc kubenswrapper[4954]: I1209 18:32:57.099269 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-utilities\") pod \"community-operators-zk2lx\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:57 crc kubenswrapper[4954]: I1209 18:32:57.137281 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxbhm\" (UniqueName: \"kubernetes.io/projected/cb9b49d8-6fc1-439a-8a16-f79936024d71-kube-api-access-vxbhm\") pod \"community-operators-zk2lx\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:57 crc kubenswrapper[4954]: I1209 18:32:57.241687 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:32:57 crc kubenswrapper[4954]: I1209 18:32:57.397261 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hlk28/crc-debug-ggzjh" event={"ID":"c789fbb9-645e-447d-aef2-b3d22c3931f3","Type":"ContainerStarted","Data":"a98a6016458c944372a49ee1c1f44b768e4fae1a15069f0c1eb02564978eb512"} Dec 09 18:32:57 crc kubenswrapper[4954]: I1209 18:32:57.876839 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zk2lx"] Dec 09 18:32:58 crc kubenswrapper[4954]: I1209 18:32:58.409491 4954 generic.go:334] "Generic (PLEG): container finished" podID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerID="4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe" exitCode=0 Dec 09 18:32:58 crc kubenswrapper[4954]: I1209 18:32:58.409575 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk2lx" event={"ID":"cb9b49d8-6fc1-439a-8a16-f79936024d71","Type":"ContainerDied","Data":"4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe"} Dec 09 18:32:58 crc kubenswrapper[4954]: I1209 18:32:58.409824 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk2lx" event={"ID":"cb9b49d8-6fc1-439a-8a16-f79936024d71","Type":"ContainerStarted","Data":"89ccfbf7a43cfba90883796cebaa743ea7e90b2f9819047e2ffae9b78b617f8f"} Dec 09 18:32:59 crc kubenswrapper[4954]: I1209 18:32:59.423280 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk2lx" event={"ID":"cb9b49d8-6fc1-439a-8a16-f79936024d71","Type":"ContainerStarted","Data":"2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041"} Dec 09 18:33:01 crc kubenswrapper[4954]: I1209 18:33:01.487391 4954 generic.go:334] "Generic (PLEG): container finished" podID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerID="2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041" exitCode=0 Dec 09 18:33:01 crc kubenswrapper[4954]: I1209 18:33:01.487458 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk2lx" event={"ID":"cb9b49d8-6fc1-439a-8a16-f79936024d71","Type":"ContainerDied","Data":"2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041"} Dec 09 18:33:02 crc kubenswrapper[4954]: I1209 18:33:02.507544 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk2lx" event={"ID":"cb9b49d8-6fc1-439a-8a16-f79936024d71","Type":"ContainerStarted","Data":"e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336"} Dec 09 18:33:02 crc kubenswrapper[4954]: I1209 18:33:02.531803 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zk2lx" podStartSLOduration=2.9669945110000002 podStartE2EDuration="6.531786706s" podCreationTimestamp="2025-12-09 18:32:56 +0000 UTC" firstStartedPulling="2025-12-09 18:32:58.411265198 +0000 UTC m=+5774.799439018" lastFinishedPulling="2025-12-09 18:33:01.976057393 +0000 UTC m=+5778.364231213" observedRunningTime="2025-12-09 18:33:02.525949733 +0000 UTC m=+5778.914123563" watchObservedRunningTime="2025-12-09 18:33:02.531786706 +0000 UTC m=+5778.919960526" Dec 09 18:33:04 crc kubenswrapper[4954]: E1209 18:33:04.131802 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:33:05 crc kubenswrapper[4954]: E1209 18:33:05.123493 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:33:07 crc kubenswrapper[4954]: I1209 18:33:07.243036 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:33:07 crc kubenswrapper[4954]: I1209 18:33:07.243639 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:33:07 crc kubenswrapper[4954]: I1209 18:33:07.297768 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:33:07 crc kubenswrapper[4954]: I1209 18:33:07.667204 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:33:07 crc kubenswrapper[4954]: I1209 18:33:07.720029 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zk2lx"] Dec 09 18:33:09 crc kubenswrapper[4954]: I1209 18:33:09.583332 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zk2lx" podUID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerName="registry-server" containerID="cri-o://e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336" gracePeriod=2 Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.178492 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.268163 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxbhm\" (UniqueName: \"kubernetes.io/projected/cb9b49d8-6fc1-439a-8a16-f79936024d71-kube-api-access-vxbhm\") pod \"cb9b49d8-6fc1-439a-8a16-f79936024d71\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.268312 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-utilities\") pod \"cb9b49d8-6fc1-439a-8a16-f79936024d71\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.268454 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-catalog-content\") pod \"cb9b49d8-6fc1-439a-8a16-f79936024d71\" (UID: \"cb9b49d8-6fc1-439a-8a16-f79936024d71\") " Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.269323 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-utilities" (OuterVolumeSpecName: "utilities") pod "cb9b49d8-6fc1-439a-8a16-f79936024d71" (UID: "cb9b49d8-6fc1-439a-8a16-f79936024d71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.281767 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb9b49d8-6fc1-439a-8a16-f79936024d71-kube-api-access-vxbhm" (OuterVolumeSpecName: "kube-api-access-vxbhm") pod "cb9b49d8-6fc1-439a-8a16-f79936024d71" (UID: "cb9b49d8-6fc1-439a-8a16-f79936024d71"). InnerVolumeSpecName "kube-api-access-vxbhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.322182 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb9b49d8-6fc1-439a-8a16-f79936024d71" (UID: "cb9b49d8-6fc1-439a-8a16-f79936024d71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.371644 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.371694 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxbhm\" (UniqueName: \"kubernetes.io/projected/cb9b49d8-6fc1-439a-8a16-f79936024d71-kube-api-access-vxbhm\") on node \"crc\" DevicePath \"\"" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.371706 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb9b49d8-6fc1-439a-8a16-f79936024d71-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.617041 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hlk28/crc-debug-ggzjh" event={"ID":"c789fbb9-645e-447d-aef2-b3d22c3931f3","Type":"ContainerStarted","Data":"4cf66fa15faeeed18d41beaa8cacba33d8bd82f4795edc2b3e300897adcb3ee0"} Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.621203 4954 generic.go:334] "Generic (PLEG): container finished" podID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerID="e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336" exitCode=0 Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.621256 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk2lx" event={"ID":"cb9b49d8-6fc1-439a-8a16-f79936024d71","Type":"ContainerDied","Data":"e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336"} Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.621307 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk2lx" event={"ID":"cb9b49d8-6fc1-439a-8a16-f79936024d71","Type":"ContainerDied","Data":"89ccfbf7a43cfba90883796cebaa743ea7e90b2f9819047e2ffae9b78b617f8f"} Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.621307 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk2lx" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.621329 4954 scope.go:117] "RemoveContainer" containerID="e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.630208 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hlk28/crc-debug-ggzjh" podStartSLOduration=2.0595373869999998 podStartE2EDuration="14.630183908s" podCreationTimestamp="2025-12-09 18:32:56 +0000 UTC" firstStartedPulling="2025-12-09 18:32:57.069716884 +0000 UTC m=+5773.457890704" lastFinishedPulling="2025-12-09 18:33:09.640363405 +0000 UTC m=+5786.028537225" observedRunningTime="2025-12-09 18:33:10.629688692 +0000 UTC m=+5787.017862522" watchObservedRunningTime="2025-12-09 18:33:10.630183908 +0000 UTC m=+5787.018357748" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.660778 4954 scope.go:117] "RemoveContainer" containerID="2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.680124 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zk2lx"] Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.685835 4954 scope.go:117] "RemoveContainer" containerID="4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.693418 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zk2lx"] Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.792773 4954 scope.go:117] "RemoveContainer" containerID="e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336" Dec 09 18:33:10 crc kubenswrapper[4954]: E1209 18:33:10.799756 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336\": container with ID starting with e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336 not found: ID does not exist" containerID="e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.799797 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336"} err="failed to get container status \"e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336\": rpc error: code = NotFound desc = could not find container \"e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336\": container with ID starting with e9ca3bc2ff73799a2e741d73d46284470b2047b1500215afb400b915f2e1b336 not found: ID does not exist" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.799822 4954 scope.go:117] "RemoveContainer" containerID="2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041" Dec 09 18:33:10 crc kubenswrapper[4954]: E1209 18:33:10.808765 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041\": container with ID starting with 2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041 not found: ID does not exist" containerID="2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.808821 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041"} err="failed to get container status \"2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041\": rpc error: code = NotFound desc = could not find container \"2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041\": container with ID starting with 2eb7da4b314c5b8a92ddbd99a8821da2abb6ee1df5c10e675f77070b892be041 not found: ID does not exist" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.808848 4954 scope.go:117] "RemoveContainer" containerID="4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe" Dec 09 18:33:10 crc kubenswrapper[4954]: E1209 18:33:10.812723 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe\": container with ID starting with 4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe not found: ID does not exist" containerID="4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe" Dec 09 18:33:10 crc kubenswrapper[4954]: I1209 18:33:10.812763 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe"} err="failed to get container status \"4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe\": rpc error: code = NotFound desc = could not find container \"4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe\": container with ID starting with 4a8a863530da2fbf59561c519658b1c8edc9c6a3424877487dda16373fda1cfe not found: ID does not exist" Dec 09 18:33:12 crc kubenswrapper[4954]: I1209 18:33:12.134098 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb9b49d8-6fc1-439a-8a16-f79936024d71" path="/var/lib/kubelet/pods/cb9b49d8-6fc1-439a-8a16-f79936024d71/volumes" Dec 09 18:33:13 crc kubenswrapper[4954]: I1209 18:33:13.755730 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:33:13 crc kubenswrapper[4954]: I1209 18:33:13.757253 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:33:13 crc kubenswrapper[4954]: I1209 18:33:13.757397 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 18:33:13 crc kubenswrapper[4954]: I1209 18:33:13.758482 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"afd4d72dcad9520c4cb369ae68cbac6019ac340837b03fa2d274871625d5a7a8"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 18:33:13 crc kubenswrapper[4954]: I1209 18:33:13.758670 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://afd4d72dcad9520c4cb369ae68cbac6019ac340837b03fa2d274871625d5a7a8" gracePeriod=600 Dec 09 18:33:14 crc kubenswrapper[4954]: I1209 18:33:14.665716 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="afd4d72dcad9520c4cb369ae68cbac6019ac340837b03fa2d274871625d5a7a8" exitCode=0 Dec 09 18:33:14 crc kubenswrapper[4954]: I1209 18:33:14.665788 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"afd4d72dcad9520c4cb369ae68cbac6019ac340837b03fa2d274871625d5a7a8"} Dec 09 18:33:14 crc kubenswrapper[4954]: I1209 18:33:14.666304 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2"} Dec 09 18:33:14 crc kubenswrapper[4954]: I1209 18:33:14.666375 4954 scope.go:117] "RemoveContainer" containerID="f3636948727fcbc46c642f4fd7c1c9e95d007f15a9c3bb8a74fd3319d53c9353" Dec 09 18:33:15 crc kubenswrapper[4954]: E1209 18:33:15.121271 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:33:17 crc kubenswrapper[4954]: E1209 18:33:17.122840 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.149242 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b5288"] Dec 09 18:33:25 crc kubenswrapper[4954]: E1209 18:33:25.151196 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerName="registry-server" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.151274 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerName="registry-server" Dec 09 18:33:25 crc kubenswrapper[4954]: E1209 18:33:25.151353 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerName="extract-utilities" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.151435 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerName="extract-utilities" Dec 09 18:33:25 crc kubenswrapper[4954]: E1209 18:33:25.151505 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerName="extract-content" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.151558 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerName="extract-content" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.151838 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb9b49d8-6fc1-439a-8a16-f79936024d71" containerName="registry-server" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.153424 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.170897 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b5288"] Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.199251 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wnq5\" (UniqueName: \"kubernetes.io/projected/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-kube-api-access-2wnq5\") pod \"certified-operators-b5288\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.199565 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-utilities\") pod \"certified-operators-b5288\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.199710 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-catalog-content\") pod \"certified-operators-b5288\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.302739 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-utilities\") pod \"certified-operators-b5288\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.302862 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-catalog-content\") pod \"certified-operators-b5288\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.303097 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wnq5\" (UniqueName: \"kubernetes.io/projected/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-kube-api-access-2wnq5\") pod \"certified-operators-b5288\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.303199 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-utilities\") pod \"certified-operators-b5288\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.303555 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-catalog-content\") pod \"certified-operators-b5288\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.338478 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wnq5\" (UniqueName: \"kubernetes.io/projected/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-kube-api-access-2wnq5\") pod \"certified-operators-b5288\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:25 crc kubenswrapper[4954]: I1209 18:33:25.470932 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:26 crc kubenswrapper[4954]: I1209 18:33:26.176199 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b5288"] Dec 09 18:33:26 crc kubenswrapper[4954]: I1209 18:33:26.846008 4954 generic.go:334] "Generic (PLEG): container finished" podID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerID="3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7" exitCode=0 Dec 09 18:33:26 crc kubenswrapper[4954]: I1209 18:33:26.846094 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5288" event={"ID":"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168","Type":"ContainerDied","Data":"3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7"} Dec 09 18:33:26 crc kubenswrapper[4954]: I1209 18:33:26.846627 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5288" event={"ID":"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168","Type":"ContainerStarted","Data":"731ffa86f13abcd410f5f76a4d297d5762f1cff9e36b95e8ff63c27e7f7e46d0"} Dec 09 18:33:26 crc kubenswrapper[4954]: I1209 18:33:26.849045 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 18:33:29 crc kubenswrapper[4954]: E1209 18:33:29.122179 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:33:29 crc kubenswrapper[4954]: I1209 18:33:29.878956 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5288" event={"ID":"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168","Type":"ContainerStarted","Data":"7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239"} Dec 09 18:33:30 crc kubenswrapper[4954]: I1209 18:33:30.890632 4954 generic.go:334] "Generic (PLEG): container finished" podID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerID="7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239" exitCode=0 Dec 09 18:33:30 crc kubenswrapper[4954]: I1209 18:33:30.890720 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5288" event={"ID":"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168","Type":"ContainerDied","Data":"7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239"} Dec 09 18:33:31 crc kubenswrapper[4954]: E1209 18:33:31.257332 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:33:31 crc kubenswrapper[4954]: E1209 18:33:31.257764 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:33:31 crc kubenswrapper[4954]: E1209 18:33:31.258014 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:33:31 crc kubenswrapper[4954]: E1209 18:33:31.259215 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:33:31 crc kubenswrapper[4954]: I1209 18:33:31.918847 4954 generic.go:334] "Generic (PLEG): container finished" podID="c789fbb9-645e-447d-aef2-b3d22c3931f3" containerID="4cf66fa15faeeed18d41beaa8cacba33d8bd82f4795edc2b3e300897adcb3ee0" exitCode=0 Dec 09 18:33:31 crc kubenswrapper[4954]: I1209 18:33:31.919042 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hlk28/crc-debug-ggzjh" event={"ID":"c789fbb9-645e-447d-aef2-b3d22c3931f3","Type":"ContainerDied","Data":"4cf66fa15faeeed18d41beaa8cacba33d8bd82f4795edc2b3e300897adcb3ee0"} Dec 09 18:33:31 crc kubenswrapper[4954]: I1209 18:33:31.931170 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5288" event={"ID":"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168","Type":"ContainerStarted","Data":"49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d"} Dec 09 18:33:31 crc kubenswrapper[4954]: I1209 18:33:31.970966 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b5288" podStartSLOduration=2.462493092 podStartE2EDuration="6.970941318s" podCreationTimestamp="2025-12-09 18:33:25 +0000 UTC" firstStartedPulling="2025-12-09 18:33:26.848715175 +0000 UTC m=+5803.236888995" lastFinishedPulling="2025-12-09 18:33:31.357163401 +0000 UTC m=+5807.745337221" observedRunningTime="2025-12-09 18:33:31.96368596 +0000 UTC m=+5808.351859800" watchObservedRunningTime="2025-12-09 18:33:31.970941318 +0000 UTC m=+5808.359115138" Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.076622 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/crc-debug-ggzjh" Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.117169 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hlk28/crc-debug-ggzjh"] Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.134835 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hlk28/crc-debug-ggzjh"] Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.187498 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txxdh\" (UniqueName: \"kubernetes.io/projected/c789fbb9-645e-447d-aef2-b3d22c3931f3-kube-api-access-txxdh\") pod \"c789fbb9-645e-447d-aef2-b3d22c3931f3\" (UID: \"c789fbb9-645e-447d-aef2-b3d22c3931f3\") " Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.187636 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c789fbb9-645e-447d-aef2-b3d22c3931f3-host\") pod \"c789fbb9-645e-447d-aef2-b3d22c3931f3\" (UID: \"c789fbb9-645e-447d-aef2-b3d22c3931f3\") " Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.187961 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c789fbb9-645e-447d-aef2-b3d22c3931f3-host" (OuterVolumeSpecName: "host") pod "c789fbb9-645e-447d-aef2-b3d22c3931f3" (UID: "c789fbb9-645e-447d-aef2-b3d22c3931f3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.188432 4954 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c789fbb9-645e-447d-aef2-b3d22c3931f3-host\") on node \"crc\" DevicePath \"\"" Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.198375 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c789fbb9-645e-447d-aef2-b3d22c3931f3-kube-api-access-txxdh" (OuterVolumeSpecName: "kube-api-access-txxdh") pod "c789fbb9-645e-447d-aef2-b3d22c3931f3" (UID: "c789fbb9-645e-447d-aef2-b3d22c3931f3"). InnerVolumeSpecName "kube-api-access-txxdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.289864 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txxdh\" (UniqueName: \"kubernetes.io/projected/c789fbb9-645e-447d-aef2-b3d22c3931f3-kube-api-access-txxdh\") on node \"crc\" DevicePath \"\"" Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.956345 4954 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a98a6016458c944372a49ee1c1f44b768e4fae1a15069f0c1eb02564978eb512" Dec 09 18:33:33 crc kubenswrapper[4954]: I1209 18:33:33.956411 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/crc-debug-ggzjh" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.132820 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c789fbb9-645e-447d-aef2-b3d22c3931f3" path="/var/lib/kubelet/pods/c789fbb9-645e-447d-aef2-b3d22c3931f3/volumes" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.368641 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hlk28/crc-debug-pvqr8"] Dec 09 18:33:34 crc kubenswrapper[4954]: E1209 18:33:34.369084 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c789fbb9-645e-447d-aef2-b3d22c3931f3" containerName="container-00" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.369100 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="c789fbb9-645e-447d-aef2-b3d22c3931f3" containerName="container-00" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.369333 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="c789fbb9-645e-447d-aef2-b3d22c3931f3" containerName="container-00" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.370053 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/crc-debug-pvqr8" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.414177 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/00c15dbc-04f7-43a4-8aaf-2adda321c720-host\") pod \"crc-debug-pvqr8\" (UID: \"00c15dbc-04f7-43a4-8aaf-2adda321c720\") " pod="openshift-must-gather-hlk28/crc-debug-pvqr8" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.414309 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bqqj\" (UniqueName: \"kubernetes.io/projected/00c15dbc-04f7-43a4-8aaf-2adda321c720-kube-api-access-5bqqj\") pod \"crc-debug-pvqr8\" (UID: \"00c15dbc-04f7-43a4-8aaf-2adda321c720\") " pod="openshift-must-gather-hlk28/crc-debug-pvqr8" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.516236 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bqqj\" (UniqueName: \"kubernetes.io/projected/00c15dbc-04f7-43a4-8aaf-2adda321c720-kube-api-access-5bqqj\") pod \"crc-debug-pvqr8\" (UID: \"00c15dbc-04f7-43a4-8aaf-2adda321c720\") " pod="openshift-must-gather-hlk28/crc-debug-pvqr8" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.516768 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/00c15dbc-04f7-43a4-8aaf-2adda321c720-host\") pod \"crc-debug-pvqr8\" (UID: \"00c15dbc-04f7-43a4-8aaf-2adda321c720\") " pod="openshift-must-gather-hlk28/crc-debug-pvqr8" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.516954 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/00c15dbc-04f7-43a4-8aaf-2adda321c720-host\") pod \"crc-debug-pvqr8\" (UID: \"00c15dbc-04f7-43a4-8aaf-2adda321c720\") " pod="openshift-must-gather-hlk28/crc-debug-pvqr8" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.541491 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bqqj\" (UniqueName: \"kubernetes.io/projected/00c15dbc-04f7-43a4-8aaf-2adda321c720-kube-api-access-5bqqj\") pod \"crc-debug-pvqr8\" (UID: \"00c15dbc-04f7-43a4-8aaf-2adda321c720\") " pod="openshift-must-gather-hlk28/crc-debug-pvqr8" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.689972 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/crc-debug-pvqr8" Dec 09 18:33:34 crc kubenswrapper[4954]: I1209 18:33:34.972826 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hlk28/crc-debug-pvqr8" event={"ID":"00c15dbc-04f7-43a4-8aaf-2adda321c720","Type":"ContainerStarted","Data":"4c68f52306813567707f733c4845f6ae35ded849de7d705c5d99599e3d9c34c3"} Dec 09 18:33:35 crc kubenswrapper[4954]: I1209 18:33:35.471280 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:35 crc kubenswrapper[4954]: I1209 18:33:35.471566 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:35 crc kubenswrapper[4954]: I1209 18:33:35.527541 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:35 crc kubenswrapper[4954]: I1209 18:33:35.984674 4954 generic.go:334] "Generic (PLEG): container finished" podID="00c15dbc-04f7-43a4-8aaf-2adda321c720" containerID="7882fe262cbd603d4b7a1fff5e4be4f75da14905ceac94d5bf24dd0c606ce376" exitCode=1 Dec 09 18:33:35 crc kubenswrapper[4954]: I1209 18:33:35.984762 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hlk28/crc-debug-pvqr8" event={"ID":"00c15dbc-04f7-43a4-8aaf-2adda321c720","Type":"ContainerDied","Data":"7882fe262cbd603d4b7a1fff5e4be4f75da14905ceac94d5bf24dd0c606ce376"} Dec 09 18:33:36 crc kubenswrapper[4954]: I1209 18:33:36.028060 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hlk28/crc-debug-pvqr8"] Dec 09 18:33:36 crc kubenswrapper[4954]: I1209 18:33:36.039874 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hlk28/crc-debug-pvqr8"] Dec 09 18:33:37 crc kubenswrapper[4954]: I1209 18:33:37.044351 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:37 crc kubenswrapper[4954]: I1209 18:33:37.106525 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/crc-debug-pvqr8" Dec 09 18:33:37 crc kubenswrapper[4954]: I1209 18:33:37.111894 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b5288"] Dec 09 18:33:37 crc kubenswrapper[4954]: I1209 18:33:37.173014 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bqqj\" (UniqueName: \"kubernetes.io/projected/00c15dbc-04f7-43a4-8aaf-2adda321c720-kube-api-access-5bqqj\") pod \"00c15dbc-04f7-43a4-8aaf-2adda321c720\" (UID: \"00c15dbc-04f7-43a4-8aaf-2adda321c720\") " Dec 09 18:33:37 crc kubenswrapper[4954]: I1209 18:33:37.173308 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/00c15dbc-04f7-43a4-8aaf-2adda321c720-host\") pod \"00c15dbc-04f7-43a4-8aaf-2adda321c720\" (UID: \"00c15dbc-04f7-43a4-8aaf-2adda321c720\") " Dec 09 18:33:37 crc kubenswrapper[4954]: I1209 18:33:37.173383 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c15dbc-04f7-43a4-8aaf-2adda321c720-host" (OuterVolumeSpecName: "host") pod "00c15dbc-04f7-43a4-8aaf-2adda321c720" (UID: "00c15dbc-04f7-43a4-8aaf-2adda321c720"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 18:33:37 crc kubenswrapper[4954]: I1209 18:33:37.173958 4954 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/00c15dbc-04f7-43a4-8aaf-2adda321c720-host\") on node \"crc\" DevicePath \"\"" Dec 09 18:33:37 crc kubenswrapper[4954]: I1209 18:33:37.179098 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00c15dbc-04f7-43a4-8aaf-2adda321c720-kube-api-access-5bqqj" (OuterVolumeSpecName: "kube-api-access-5bqqj") pod "00c15dbc-04f7-43a4-8aaf-2adda321c720" (UID: "00c15dbc-04f7-43a4-8aaf-2adda321c720"). InnerVolumeSpecName "kube-api-access-5bqqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:33:37 crc kubenswrapper[4954]: I1209 18:33:37.275299 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bqqj\" (UniqueName: \"kubernetes.io/projected/00c15dbc-04f7-43a4-8aaf-2adda321c720-kube-api-access-5bqqj\") on node \"crc\" DevicePath \"\"" Dec 09 18:33:38 crc kubenswrapper[4954]: I1209 18:33:38.016367 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/crc-debug-pvqr8" Dec 09 18:33:38 crc kubenswrapper[4954]: I1209 18:33:38.016404 4954 scope.go:117] "RemoveContainer" containerID="7882fe262cbd603d4b7a1fff5e4be4f75da14905ceac94d5bf24dd0c606ce376" Dec 09 18:33:38 crc kubenswrapper[4954]: I1209 18:33:38.132138 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00c15dbc-04f7-43a4-8aaf-2adda321c720" path="/var/lib/kubelet/pods/00c15dbc-04f7-43a4-8aaf-2adda321c720/volumes" Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.026793 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b5288" podUID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerName="registry-server" containerID="cri-o://49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d" gracePeriod=2 Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.561739 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.723278 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-utilities\") pod \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.723408 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-catalog-content\") pod \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.723502 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wnq5\" (UniqueName: \"kubernetes.io/projected/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-kube-api-access-2wnq5\") pod \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\" (UID: \"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168\") " Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.724117 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-utilities" (OuterVolumeSpecName: "utilities") pod "b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" (UID: "b684cdd8-1c8e-4a1f-9b0c-98678ffe4168"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.735394 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-kube-api-access-2wnq5" (OuterVolumeSpecName: "kube-api-access-2wnq5") pod "b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" (UID: "b684cdd8-1c8e-4a1f-9b0c-98678ffe4168"). InnerVolumeSpecName "kube-api-access-2wnq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.771659 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" (UID: "b684cdd8-1c8e-4a1f-9b0c-98678ffe4168"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.825984 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.826017 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:33:39 crc kubenswrapper[4954]: I1209 18:33:39.826029 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wnq5\" (UniqueName: \"kubernetes.io/projected/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168-kube-api-access-2wnq5\") on node \"crc\" DevicePath \"\"" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.038057 4954 generic.go:334] "Generic (PLEG): container finished" podID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerID="49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d" exitCode=0 Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.038112 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5288" event={"ID":"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168","Type":"ContainerDied","Data":"49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d"} Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.038146 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5288" event={"ID":"b684cdd8-1c8e-4a1f-9b0c-98678ffe4168","Type":"ContainerDied","Data":"731ffa86f13abcd410f5f76a4d297d5762f1cff9e36b95e8ff63c27e7f7e46d0"} Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.038169 4954 scope.go:117] "RemoveContainer" containerID="49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.038336 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5288" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.063553 4954 scope.go:117] "RemoveContainer" containerID="7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.081755 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b5288"] Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.099278 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b5288"] Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.105246 4954 scope.go:117] "RemoveContainer" containerID="3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.151310 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" path="/var/lib/kubelet/pods/b684cdd8-1c8e-4a1f-9b0c-98678ffe4168/volumes" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.177768 4954 scope.go:117] "RemoveContainer" containerID="49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d" Dec 09 18:33:40 crc kubenswrapper[4954]: E1209 18:33:40.178319 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d\": container with ID starting with 49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d not found: ID does not exist" containerID="49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.178437 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d"} err="failed to get container status \"49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d\": rpc error: code = NotFound desc = could not find container \"49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d\": container with ID starting with 49abfdb17e61b28901b15bbfbaabeda7ba232d528aff0d0ccfcbbb1965db664d not found: ID does not exist" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.178520 4954 scope.go:117] "RemoveContainer" containerID="7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239" Dec 09 18:33:40 crc kubenswrapper[4954]: E1209 18:33:40.179128 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239\": container with ID starting with 7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239 not found: ID does not exist" containerID="7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.179187 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239"} err="failed to get container status \"7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239\": rpc error: code = NotFound desc = could not find container \"7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239\": container with ID starting with 7a627b641b54290781f39f7bb1de8b553d1830d7e18b5b4be59429c98e096239 not found: ID does not exist" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.179223 4954 scope.go:117] "RemoveContainer" containerID="3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7" Dec 09 18:33:40 crc kubenswrapper[4954]: E1209 18:33:40.179488 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7\": container with ID starting with 3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7 not found: ID does not exist" containerID="3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7" Dec 09 18:33:40 crc kubenswrapper[4954]: I1209 18:33:40.179575 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7"} err="failed to get container status \"3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7\": rpc error: code = NotFound desc = could not find container \"3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7\": container with ID starting with 3201e2e40cfd024d2ca3db4b602558863db8ab0f858405473584a1dba8a4c7d7 not found: ID does not exist" Dec 09 18:33:42 crc kubenswrapper[4954]: E1209 18:33:42.123225 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:33:42 crc kubenswrapper[4954]: E1209 18:33:42.123228 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:33:53 crc kubenswrapper[4954]: E1209 18:33:53.125323 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:33:56 crc kubenswrapper[4954]: E1209 18:33:56.248719 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:33:56 crc kubenswrapper[4954]: E1209 18:33:56.249019 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:33:56 crc kubenswrapper[4954]: E1209 18:33:56.249145 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:33:56 crc kubenswrapper[4954]: E1209 18:33:56.251001 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:34:05 crc kubenswrapper[4954]: E1209 18:34:05.125133 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:34:09 crc kubenswrapper[4954]: E1209 18:34:09.123420 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:34:19 crc kubenswrapper[4954]: E1209 18:34:19.123074 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:34:23 crc kubenswrapper[4954]: I1209 18:34:23.350386 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e2935140-950a-4c06-adee-a72196dec3d8/aodh-api/0.log" Dec 09 18:34:23 crc kubenswrapper[4954]: I1209 18:34:23.544080 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e2935140-950a-4c06-adee-a72196dec3d8/aodh-notifier/0.log" Dec 09 18:34:23 crc kubenswrapper[4954]: I1209 18:34:23.544398 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e2935140-950a-4c06-adee-a72196dec3d8/aodh-listener/0.log" Dec 09 18:34:23 crc kubenswrapper[4954]: I1209 18:34:23.552714 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e2935140-950a-4c06-adee-a72196dec3d8/aodh-evaluator/0.log" Dec 09 18:34:23 crc kubenswrapper[4954]: I1209 18:34:23.720583 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dfdb467d-r2rqf_653b2aa3-1e55-42ae-9284-51d651461a15/barbican-api/0.log" Dec 09 18:34:23 crc kubenswrapper[4954]: I1209 18:34:23.739324 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6dfdb467d-r2rqf_653b2aa3-1e55-42ae-9284-51d651461a15/barbican-api-log/0.log" Dec 09 18:34:23 crc kubenswrapper[4954]: I1209 18:34:23.884898 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7cbfd68ff6-49k8v_ad27e7b2-2f17-4dbc-b611-53f192f502ae/barbican-keystone-listener/0.log" Dec 09 18:34:23 crc kubenswrapper[4954]: I1209 18:34:23.923771 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7cbfd68ff6-49k8v_ad27e7b2-2f17-4dbc-b611-53f192f502ae/barbican-keystone-listener-log/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.096814 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76bccc865f-nls4t_704e4510-1e1b-492f-9de5-e5bbec8e6731/barbican-worker/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.124169 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76bccc865f-nls4t_704e4510-1e1b-492f-9de5-e5bbec8e6731/barbican-worker-log/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: E1209 18:34:24.130959 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.195546 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-psvvd_43bffa1b-06d5-470d-b410-82f6acf14286/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.399264 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df/ceilometer-notification-agent/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.430129 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df/proxy-httpd/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.555402 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-v9jt7_1062cfab-9c0f-4965-a8bf-75ff7f1744d2/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.589097 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df/sg-core/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.779372 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a8078c26-0c70-430b-b3d8-6f56d236657c/cinder-api/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.787226 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a8078c26-0c70-430b-b3d8-6f56d236657c/cinder-api-log/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.880623 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_74bc3bd9-6b4e-435c-b82e-3b1378962d6a/cinder-scheduler/0.log" Dec 09 18:34:24 crc kubenswrapper[4954]: I1209 18:34:24.994381 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_74bc3bd9-6b4e-435c-b82e-3b1378962d6a/probe/0.log" Dec 09 18:34:25 crc kubenswrapper[4954]: I1209 18:34:25.050299 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-ppqwv_dbd5865d-2c5b-415f-a824-d0acfefad34a/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:25 crc kubenswrapper[4954]: I1209 18:34:25.253642 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-v7cn8_b6a4b196-e34d-4b20-9915-f290a4dac804/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:25 crc kubenswrapper[4954]: I1209 18:34:25.310059 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6559847fc9-v6xcs_c37d726c-59ef-4dd7-8a00-32417dfe1a19/init/0.log" Dec 09 18:34:25 crc kubenswrapper[4954]: I1209 18:34:25.424136 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6559847fc9-v6xcs_c37d726c-59ef-4dd7-8a00-32417dfe1a19/init/0.log" Dec 09 18:34:25 crc kubenswrapper[4954]: I1209 18:34:25.464207 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6559847fc9-v6xcs_c37d726c-59ef-4dd7-8a00-32417dfe1a19/dnsmasq-dns/0.log" Dec 09 18:34:26 crc kubenswrapper[4954]: I1209 18:34:26.026382 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-6744cc5f97-xg7zz_03ea0f62-7526-4cc9-905a-cfd0d5f791e0/heat-api/0.log" Dec 09 18:34:26 crc kubenswrapper[4954]: I1209 18:34:26.239678 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-7776d8d5fb-j42q6_1241e34b-a815-4395-8c78-16b79f6e8ecb/heat-cfnapi/0.log" Dec 09 18:34:26 crc kubenswrapper[4954]: I1209 18:34:26.312184 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-6d888884cc-94rdv_5d0a68d8-91d2-4453-8a36-6e4616a88219/heat-engine/0.log" Dec 09 18:34:26 crc kubenswrapper[4954]: I1209 18:34:26.545780 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-gs5fd_2251824f-c169-44ff-afeb-8649fb57c39f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:26 crc kubenswrapper[4954]: I1209 18:34:26.599182 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-mg9sk_511ba2be-4b0f-4d57-8ff8-d2a25f541fd3/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:26 crc kubenswrapper[4954]: I1209 18:34:26.850334 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-64ffb74c6f-59w54_4e460517-3e4e-4179-b4f0-1b1db948715b/keystone-api/0.log" Dec 09 18:34:26 crc kubenswrapper[4954]: I1209 18:34:26.855653 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29421721-b7lcm_71a35b4e-c2ab-4268-b0a2-2e5d83e54d9d/keystone-cron/0.log" Dec 09 18:34:27 crc kubenswrapper[4954]: I1209 18:34:27.066852 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_03899af3-5963-4f63-8c51-8fcbbc8582f0/kube-state-metrics/0.log" Dec 09 18:34:27 crc kubenswrapper[4954]: I1209 18:34:27.138202 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-wpc4k_d64c4820-0dfe-46fa-9275-e9d45bbb268e/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:27 crc kubenswrapper[4954]: I1209 18:34:27.334772 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_9785110a-93dc-4330-8198-792183afdb87/mysqld-exporter/0.log" Dec 09 18:34:27 crc kubenswrapper[4954]: I1209 18:34:27.575974 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9855f74b5-lqsx6_e52dc021-9970-4cec-8e66-beb3b49a1dd5/neutron-httpd/0.log" Dec 09 18:34:27 crc kubenswrapper[4954]: I1209 18:34:27.638112 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9855f74b5-lqsx6_e52dc021-9970-4cec-8e66-beb3b49a1dd5/neutron-api/0.log" Dec 09 18:34:27 crc kubenswrapper[4954]: I1209 18:34:27.974488 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_01db8384-81a2-4c9d-b26d-a8d788614768/nova-api-log/0.log" Dec 09 18:34:28 crc kubenswrapper[4954]: I1209 18:34:28.028489 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_5f3b535e-6a1e-4ff0-abdf-be624309bb19/nova-cell0-conductor-conductor/0.log" Dec 09 18:34:28 crc kubenswrapper[4954]: I1209 18:34:28.285358 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_01db8384-81a2-4c9d-b26d-a8d788614768/nova-api-api/0.log" Dec 09 18:34:28 crc kubenswrapper[4954]: I1209 18:34:28.312341 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_cf11f1d4-587b-49a1-bd96-153799d2a6d4/nova-cell1-conductor-conductor/0.log" Dec 09 18:34:28 crc kubenswrapper[4954]: I1209 18:34:28.476969 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_29eaf432-7585-424b-9732-f7e0bb49f380/nova-cell1-novncproxy-novncproxy/0.log" Dec 09 18:34:28 crc kubenswrapper[4954]: I1209 18:34:28.629914 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f299ddad-e1de-4ea5-a76c-35477739f14e/nova-metadata-log/0.log" Dec 09 18:34:28 crc kubenswrapper[4954]: I1209 18:34:28.927048 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_7cb9e244-80bc-4b7b-8b89-c44bdb73ae80/nova-scheduler-scheduler/0.log" Dec 09 18:34:29 crc kubenswrapper[4954]: I1209 18:34:29.008013 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_197b01d4-b9ac-4dc6-ac33-4638ece0241f/mysql-bootstrap/0.log" Dec 09 18:34:29 crc kubenswrapper[4954]: I1209 18:34:29.233890 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_197b01d4-b9ac-4dc6-ac33-4638ece0241f/mysql-bootstrap/0.log" Dec 09 18:34:29 crc kubenswrapper[4954]: I1209 18:34:29.288433 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_197b01d4-b9ac-4dc6-ac33-4638ece0241f/galera/0.log" Dec 09 18:34:29 crc kubenswrapper[4954]: I1209 18:34:29.488855 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d42e4546-9029-4930-ab04-7adaab383d5a/mysql-bootstrap/0.log" Dec 09 18:34:29 crc kubenswrapper[4954]: I1209 18:34:29.717166 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d42e4546-9029-4930-ab04-7adaab383d5a/mysql-bootstrap/0.log" Dec 09 18:34:29 crc kubenswrapper[4954]: I1209 18:34:29.730240 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d42e4546-9029-4930-ab04-7adaab383d5a/galera/0.log" Dec 09 18:34:29 crc kubenswrapper[4954]: I1209 18:34:29.960266 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6f8fc75c-fdd0-4931-9e61-02bccaee6b6d/openstackclient/0.log" Dec 09 18:34:29 crc kubenswrapper[4954]: I1209 18:34:29.990378 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-l6s5g_94a5a72d-c7dd-462f-969f-812555bc9ba3/ovn-controller/0.log" Dec 09 18:34:30 crc kubenswrapper[4954]: I1209 18:34:30.130898 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-g9rnm_27f44545-77d2-4048-9bfc-cd8504a1b837/openstack-network-exporter/0.log" Dec 09 18:34:30 crc kubenswrapper[4954]: I1209 18:34:30.586181 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ggbdh_2352a5dd-0445-4b59-9e8e-4607d055e505/ovsdb-server-init/0.log" Dec 09 18:34:30 crc kubenswrapper[4954]: I1209 18:34:30.596832 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f299ddad-e1de-4ea5-a76c-35477739f14e/nova-metadata-metadata/0.log" Dec 09 18:34:30 crc kubenswrapper[4954]: I1209 18:34:30.823283 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ggbdh_2352a5dd-0445-4b59-9e8e-4607d055e505/ovsdb-server/0.log" Dec 09 18:34:30 crc kubenswrapper[4954]: I1209 18:34:30.836612 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ggbdh_2352a5dd-0445-4b59-9e8e-4607d055e505/ovs-vswitchd/0.log" Dec 09 18:34:30 crc kubenswrapper[4954]: I1209 18:34:30.854764 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ggbdh_2352a5dd-0445-4b59-9e8e-4607d055e505/ovsdb-server-init/0.log" Dec 09 18:34:31 crc kubenswrapper[4954]: I1209 18:34:31.068582 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-jt7gl_a20b6b58-cb72-432d-8ec8-7c7f93265d36/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:31 crc kubenswrapper[4954]: I1209 18:34:31.125156 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b77bf96d-5422-416f-ba9d-cd20e94d991e/openstack-network-exporter/0.log" Dec 09 18:34:31 crc kubenswrapper[4954]: I1209 18:34:31.184076 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b77bf96d-5422-416f-ba9d-cd20e94d991e/ovn-northd/0.log" Dec 09 18:34:31 crc kubenswrapper[4954]: I1209 18:34:31.283033 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_060bba41-7ec2-4056-9d55-7b9d4ea22b96/openstack-network-exporter/0.log" Dec 09 18:34:31 crc kubenswrapper[4954]: I1209 18:34:31.349441 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_060bba41-7ec2-4056-9d55-7b9d4ea22b96/ovsdbserver-nb/0.log" Dec 09 18:34:31 crc kubenswrapper[4954]: I1209 18:34:31.475584 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_619c2101-1cbd-4584-a5fd-b72f9963b9d6/openstack-network-exporter/0.log" Dec 09 18:34:31 crc kubenswrapper[4954]: I1209 18:34:31.637420 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_619c2101-1cbd-4584-a5fd-b72f9963b9d6/ovsdbserver-sb/0.log" Dec 09 18:34:31 crc kubenswrapper[4954]: I1209 18:34:31.751335 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-57bb47ccd8-dn5h5_ce0dd183-082d-42f3-8e74-050349e431fa/placement-api/0.log" Dec 09 18:34:31 crc kubenswrapper[4954]: I1209 18:34:31.793399 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-57bb47ccd8-dn5h5_ce0dd183-082d-42f3-8e74-050349e431fa/placement-log/0.log" Dec 09 18:34:31 crc kubenswrapper[4954]: I1209 18:34:31.945467 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e626f486-ed0d-4388-9040-1c23a41aba18/init-config-reloader/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.132213 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e626f486-ed0d-4388-9040-1c23a41aba18/init-config-reloader/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.132408 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e626f486-ed0d-4388-9040-1c23a41aba18/config-reloader/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.167712 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e626f486-ed0d-4388-9040-1c23a41aba18/prometheus/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.182787 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e626f486-ed0d-4388-9040-1c23a41aba18/thanos-sidecar/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.324427 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fa9d388c-71e6-4332-9e32-de049db360e1/setup-container/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.660889 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fa9d388c-71e6-4332-9e32-de049db360e1/rabbitmq/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.683100 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fa9d388c-71e6-4332-9e32-de049db360e1/setup-container/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.722803 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ccb386f6-c327-4832-a461-7dd5cb0b44b7/setup-container/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.863386 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ccb386f6-c327-4832-a461-7dd5cb0b44b7/setup-container/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.976018 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ccb386f6-c327-4832-a461-7dd5cb0b44b7/rabbitmq/0.log" Dec 09 18:34:32 crc kubenswrapper[4954]: I1209 18:34:32.982556 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-chvp8_a7681334-f1d0-409b-a4e0-742786731cda/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:33 crc kubenswrapper[4954]: E1209 18:34:33.123541 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:34:33 crc kubenswrapper[4954]: I1209 18:34:33.251808 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-qzr2s_fc8a05c8-a7d1-440d-b800-f0a4217ec7f6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:33 crc kubenswrapper[4954]: I1209 18:34:33.325940 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-rvnx4_f7031e05-5441-47fa-84f5-f35fada4a4ba/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:33 crc kubenswrapper[4954]: I1209 18:34:33.528435 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-2w25t_b82e09bb-ec3a-4aa9-ba9e-ef0bcbb42dd1/ssh-known-hosts-edpm-deployment/0.log" Dec 09 18:34:33 crc kubenswrapper[4954]: I1209 18:34:33.690042 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-547f8767c7-hv854_a6635c8e-d9a5-4034-ab5c-7fe96dc10c10/proxy-server/0.log" Dec 09 18:34:33 crc kubenswrapper[4954]: I1209 18:34:33.847655 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-547f8767c7-hv854_a6635c8e-d9a5-4034-ab5c-7fe96dc10c10/proxy-httpd/0.log" Dec 09 18:34:33 crc kubenswrapper[4954]: I1209 18:34:33.847735 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-4vtb6_9fc44917-47c6-450b-b393-bad021024fcd/swift-ring-rebalance/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.024608 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/account-auditor/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.080832 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/account-reaper/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.137138 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/account-replicator/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.171503 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/account-server/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.288587 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/container-auditor/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.348052 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/container-replicator/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.351143 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/container-server/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.445769 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/container-updater/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.669703 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/object-expirer/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.684558 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/object-auditor/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.738961 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/object-replicator/0.log" Dec 09 18:34:34 crc kubenswrapper[4954]: I1209 18:34:34.925508 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/object-server/0.log" Dec 09 18:34:35 crc kubenswrapper[4954]: I1209 18:34:35.014104 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/object-updater/0.log" Dec 09 18:34:35 crc kubenswrapper[4954]: I1209 18:34:35.086382 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/swift-recon-cron/0.log" Dec 09 18:34:35 crc kubenswrapper[4954]: I1209 18:34:35.164391 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69b248e7-6cb4-4805-a01f-537fe58aa42d/rsync/0.log" Dec 09 18:34:35 crc kubenswrapper[4954]: I1209 18:34:35.300043 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-5m8pp_ac8da3e6-984f-4813-bd51-595981f71106/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:35 crc kubenswrapper[4954]: I1209 18:34:35.445847 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-6lr78_b21d50b3-8256-4f9e-ad87-681ff4111133/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:35 crc kubenswrapper[4954]: I1209 18:34:35.521111 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-7z7tq_9d2dcb4c-f20a-467a-9e39-4577f464c3f1/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:35 crc kubenswrapper[4954]: I1209 18:34:35.693165 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-87vhd_31ae7df8-13b6-4691-8d2b-3b5880921353/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:35 crc kubenswrapper[4954]: I1209 18:34:35.779377 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-lkcbm_5aace7ec-d1b0-4c50-890a-9c83e3e530fc/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:35 crc kubenswrapper[4954]: I1209 18:34:35.950044 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-w7sxd_1eb219ed-3f84-41ec-be8c-1e5d5f8c7dfc/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:36 crc kubenswrapper[4954]: I1209 18:34:36.065138 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-z4rjl_0575ea51-efad-4a28-9cf1-b2b2653312d1/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:36 crc kubenswrapper[4954]: E1209 18:34:36.122279 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:34:36 crc kubenswrapper[4954]: I1209 18:34:36.230983 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-cplh4_5f717719-b2db-4a33-9607-dc210c19261f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 18:34:40 crc kubenswrapper[4954]: I1209 18:34:40.604149 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_bb8de77c-60eb-4e0d-abdf-d6bc9f111031/memcached/0.log" Dec 09 18:34:48 crc kubenswrapper[4954]: E1209 18:34:48.123920 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:34:49 crc kubenswrapper[4954]: E1209 18:34:49.121569 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:34:59 crc kubenswrapper[4954]: E1209 18:34:59.123298 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:35:00 crc kubenswrapper[4954]: E1209 18:35:00.123685 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.024828 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b_d585c091-cf8b-41ef-8f4f-bfdfb004fd10/util/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.186361 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b_d585c091-cf8b-41ef-8f4f-bfdfb004fd10/util/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.186813 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b_d585c091-cf8b-41ef-8f4f-bfdfb004fd10/pull/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.232513 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b_d585c091-cf8b-41ef-8f4f-bfdfb004fd10/pull/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.379564 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b_d585c091-cf8b-41ef-8f4f-bfdfb004fd10/util/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.387670 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b_d585c091-cf8b-41ef-8f4f-bfdfb004fd10/pull/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.452677 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_84f5b8ce9b737d136e9d44026726099c3c124bf2a1b3be498eb888ce47lxz2b_d585c091-cf8b-41ef-8f4f-bfdfb004fd10/extract/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.580727 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4pcjs_e57eccbb-b484-45ad-a474-292da5bb6547/kube-rbac-proxy/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.674293 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-6sfrq_54373352-1448-443c-82eb-4bd28627acdf/kube-rbac-proxy/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.681461 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4pcjs_e57eccbb-b484-45ad-a474-292da5bb6547/manager/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.816432 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-6sfrq_54373352-1448-443c-82eb-4bd28627acdf/manager/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.845951 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-8857b_41401dcd-a6d4-4e4b-93e9-bea5977d7e90/kube-rbac-proxy/0.log" Dec 09 18:35:02 crc kubenswrapper[4954]: I1209 18:35:02.882495 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-8857b_41401dcd-a6d4-4e4b-93e9-bea5977d7e90/manager/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.018347 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-k5vlj_3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8/kube-rbac-proxy/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.090335 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-k5vlj_3d7b0ce5-bf71-4681-a92a-ee6c4b76a9e8/manager/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.239931 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-8mpbt_917a4c41-4cec-49ef-95f7-6cda64da9c44/kube-rbac-proxy/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.338690 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-28bcg_74b54335-7821-4cdc-a994-1dbe6c1653c9/kube-rbac-proxy/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.388786 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-8mpbt_917a4c41-4cec-49ef-95f7-6cda64da9c44/manager/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.421628 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-28bcg_74b54335-7821-4cdc-a994-1dbe6c1653c9/manager/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.570095 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-cdpd9_2ec773b3-dd1d-4662-9ebb-4be2993f76a2/kube-rbac-proxy/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.790853 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-hlx7x_2b64512f-8bdc-42b5-9269-5ea7cd1eabf1/kube-rbac-proxy/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.850579 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-cdpd9_2ec773b3-dd1d-4662-9ebb-4be2993f76a2/manager/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.862674 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-hlx7x_2b64512f-8bdc-42b5-9269-5ea7cd1eabf1/manager/0.log" Dec 09 18:35:03 crc kubenswrapper[4954]: I1209 18:35:03.983656 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-76tb6_e84bf071-0cf4-4e92-9060-6cc7d30984db/kube-rbac-proxy/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.087178 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-76tb6_e84bf071-0cf4-4e92-9060-6cc7d30984db/manager/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.155172 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-g6ggb_6000ab00-d760-4e35-9ef6-47b878a8ef96/kube-rbac-proxy/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.185012 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-g6ggb_6000ab00-d760-4e35-9ef6-47b878a8ef96/manager/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.298282 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-pfvmf_8f82f473-cbee-4499-89b6-e6fbc32405aa/kube-rbac-proxy/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.377956 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-pfvmf_8f82f473-cbee-4499-89b6-e6fbc32405aa/manager/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.505908 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-qsdpz_37d5d0b5-1e21-4b84-bd10-f4f738ef37a1/kube-rbac-proxy/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.586314 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-qsdpz_37d5d0b5-1e21-4b84-bd10-f4f738ef37a1/manager/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.590426 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-mz7gh_dc3d4dbd-0b2e-489e-b686-0195c372a917/kube-rbac-proxy/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.779951 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-5jjz9_2ba970e1-21a2-477e-947f-89de7bfbb6a6/kube-rbac-proxy/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.785105 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-mz7gh_dc3d4dbd-0b2e-489e-b686-0195c372a917/manager/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.789854 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-5jjz9_2ba970e1-21a2-477e-947f-89de7bfbb6a6/manager/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.941431 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fzcn6j_074bde7a-88ab-45ac-8a29-09f49834e4c8/kube-rbac-proxy/0.log" Dec 09 18:35:04 crc kubenswrapper[4954]: I1209 18:35:04.975274 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fzcn6j_074bde7a-88ab-45ac-8a29-09f49834e4c8/manager/0.log" Dec 09 18:35:05 crc kubenswrapper[4954]: I1209 18:35:05.354031 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5f5557f974-fsx98_4a36b355-e0bf-4194-8c6a-fe39d7fa4b16/operator/0.log" Dec 09 18:35:05 crc kubenswrapper[4954]: I1209 18:35:05.385378 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-5t29f_428c0773-ebc5-4b31-b240-7aa8f6c2fcc9/registry-server/0.log" Dec 09 18:35:05 crc kubenswrapper[4954]: I1209 18:35:05.573375 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-n7kmf_281087c4-abfa-4537-80b5-130b03885954/kube-rbac-proxy/0.log" Dec 09 18:35:05 crc kubenswrapper[4954]: I1209 18:35:05.692423 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-n7kmf_281087c4-abfa-4537-80b5-130b03885954/manager/0.log" Dec 09 18:35:05 crc kubenswrapper[4954]: I1209 18:35:05.748723 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nvdgc_410c21d8-1970-411a-8704-db44a36c74cd/kube-rbac-proxy/0.log" Dec 09 18:35:05 crc kubenswrapper[4954]: I1209 18:35:05.963833 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nvdgc_410c21d8-1970-411a-8704-db44a36c74cd/manager/0.log" Dec 09 18:35:05 crc kubenswrapper[4954]: I1209 18:35:05.993960 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-rcwxm_1fc4c8d7-5079-45d0-8a68-d2523e7b981f/operator/0.log" Dec 09 18:35:06 crc kubenswrapper[4954]: I1209 18:35:06.226501 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-8ltxl_3c0851fe-4209-4e7f-9622-dba385e6bc78/kube-rbac-proxy/0.log" Dec 09 18:35:06 crc kubenswrapper[4954]: I1209 18:35:06.228276 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-8ltxl_3c0851fe-4209-4e7f-9622-dba385e6bc78/manager/0.log" Dec 09 18:35:06 crc kubenswrapper[4954]: I1209 18:35:06.344818 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-796785f986-j59bx_93405692-e3d5-4a81-97db-170d70b6cbc3/kube-rbac-proxy/0.log" Dec 09 18:35:06 crc kubenswrapper[4954]: I1209 18:35:06.528230 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-866b78c4d6-m6k5w_4ec17be0-419e-48e3-a00f-58741723b85d/manager/0.log" Dec 09 18:35:06 crc kubenswrapper[4954]: I1209 18:35:06.548705 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-96qpl_ce9db630-dee1-4c9f-b90e-31d263b8a213/kube-rbac-proxy/0.log" Dec 09 18:35:06 crc kubenswrapper[4954]: I1209 18:35:06.585497 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-96qpl_ce9db630-dee1-4c9f-b90e-31d263b8a213/manager/0.log" Dec 09 18:35:06 crc kubenswrapper[4954]: I1209 18:35:06.743026 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-796785f986-j59bx_93405692-e3d5-4a81-97db-170d70b6cbc3/manager/0.log" Dec 09 18:35:06 crc kubenswrapper[4954]: I1209 18:35:06.749533 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-4rrqn_1f87c2e5-6b6b-4f67-b2f6-caf321fbe844/kube-rbac-proxy/0.log" Dec 09 18:35:06 crc kubenswrapper[4954]: I1209 18:35:06.757897 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-4rrqn_1f87c2e5-6b6b-4f67-b2f6-caf321fbe844/manager/0.log" Dec 09 18:35:13 crc kubenswrapper[4954]: E1209 18:35:13.123505 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:35:14 crc kubenswrapper[4954]: E1209 18:35:14.131718 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:35:25 crc kubenswrapper[4954]: E1209 18:35:25.123386 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:35:26 crc kubenswrapper[4954]: E1209 18:35:26.122354 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:35:26 crc kubenswrapper[4954]: I1209 18:35:26.408109 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-7jpc4_8f80de45-61f5-4538-93d7-c6252e3c64bc/control-plane-machine-set-operator/0.log" Dec 09 18:35:26 crc kubenswrapper[4954]: I1209 18:35:26.606701 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h7bjt_8ffdbd23-8e68-4bf3-bb37-2817fa11f18b/kube-rbac-proxy/0.log" Dec 09 18:35:26 crc kubenswrapper[4954]: I1209 18:35:26.625101 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h7bjt_8ffdbd23-8e68-4bf3-bb37-2817fa11f18b/machine-api-operator/0.log" Dec 09 18:35:36 crc kubenswrapper[4954]: E1209 18:35:36.122357 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:35:39 crc kubenswrapper[4954]: I1209 18:35:39.039362 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-pj9hg_f09a1e60-44f9-4ce4-b136-ab1131433933/cert-manager-controller/0.log" Dec 09 18:35:39 crc kubenswrapper[4954]: E1209 18:35:39.129829 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:35:39 crc kubenswrapper[4954]: I1209 18:35:39.334276 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-lwjmk_19d20d7b-da5f-446f-b2f8-cf6ab0864461/cert-manager-webhook/0.log" Dec 09 18:35:39 crc kubenswrapper[4954]: I1209 18:35:39.344912 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-ltkfg_cec94679-b7b3-4210-9c39-a4f2c012266c/cert-manager-cainjector/0.log" Dec 09 18:35:43 crc kubenswrapper[4954]: I1209 18:35:43.755824 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:35:43 crc kubenswrapper[4954]: I1209 18:35:43.756442 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:35:51 crc kubenswrapper[4954]: E1209 18:35:51.122966 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:35:52 crc kubenswrapper[4954]: E1209 18:35:52.122176 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:35:52 crc kubenswrapper[4954]: I1209 18:35:52.203816 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-tvljd_17ebcbf9-bb08-4a4f-b0c6-e081049376cf/nmstate-console-plugin/0.log" Dec 09 18:35:52 crc kubenswrapper[4954]: I1209 18:35:52.389658 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-r7xlm_a3a905e4-153e-46de-879b-45e0870277b4/kube-rbac-proxy/0.log" Dec 09 18:35:52 crc kubenswrapper[4954]: I1209 18:35:52.399006 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-k85lr_8b54c655-482f-4cfa-8029-c9fd0da67d17/nmstate-handler/0.log" Dec 09 18:35:52 crc kubenswrapper[4954]: I1209 18:35:52.438737 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-r7xlm_a3a905e4-153e-46de-879b-45e0870277b4/nmstate-metrics/0.log" Dec 09 18:35:52 crc kubenswrapper[4954]: I1209 18:35:52.638822 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-rls9q_71dff39f-cf61-434a-9c53-ad96e18d363f/nmstate-webhook/0.log" Dec 09 18:35:52 crc kubenswrapper[4954]: I1209 18:35:52.640359 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-682nr_f365441f-c01b-4f49-834c-30326629074e/nmstate-operator/0.log" Dec 09 18:36:05 crc kubenswrapper[4954]: I1209 18:36:05.680247 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5f76df48c6-fvg95_73f81f6d-15c4-4cc8-95be-59c4db2139ea/manager/0.log" Dec 09 18:36:05 crc kubenswrapper[4954]: I1209 18:36:05.689617 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5f76df48c6-fvg95_73f81f6d-15c4-4cc8-95be-59c4db2139ea/kube-rbac-proxy/0.log" Dec 09 18:36:06 crc kubenswrapper[4954]: E1209 18:36:06.124575 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:36:07 crc kubenswrapper[4954]: E1209 18:36:07.121447 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:36:13 crc kubenswrapper[4954]: I1209 18:36:13.755682 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:36:13 crc kubenswrapper[4954]: I1209 18:36:13.756166 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:36:19 crc kubenswrapper[4954]: E1209 18:36:19.122528 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:36:19 crc kubenswrapper[4954]: I1209 18:36:19.897999 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-c7nbc_0f07fc00-40fe-4979-9225-d66d3c359c2c/cluster-logging-operator/0.log" Dec 09 18:36:20 crc kubenswrapper[4954]: I1209 18:36:20.080671 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-87lsn_7302a86b-a768-46f5-af6b-e3429564341d/collector/0.log" Dec 09 18:36:20 crc kubenswrapper[4954]: I1209 18:36:20.205917 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_7fd1add7-4d51-4474-a1c4-b95df89c1bb1/loki-compactor/0.log" Dec 09 18:36:20 crc kubenswrapper[4954]: I1209 18:36:20.280473 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-hgwvr_8ca5d9e5-51f2-4a7a-86e4-a862190a0828/loki-distributor/0.log" Dec 09 18:36:20 crc kubenswrapper[4954]: I1209 18:36:20.390045 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-676969bccf-hqhwj_49d8d406-58db-4226-806b-850e6e12b04f/gateway/0.log" Dec 09 18:36:20 crc kubenswrapper[4954]: I1209 18:36:20.492260 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-676969bccf-hqhwj_49d8d406-58db-4226-806b-850e6e12b04f/opa/0.log" Dec 09 18:36:20 crc kubenswrapper[4954]: I1209 18:36:20.610026 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-676969bccf-mxr7n_81042cee-7dc6-4fc9-b8cf-70a4eff73510/gateway/0.log" Dec 09 18:36:20 crc kubenswrapper[4954]: I1209 18:36:20.612332 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-676969bccf-mxr7n_81042cee-7dc6-4fc9-b8cf-70a4eff73510/opa/0.log" Dec 09 18:36:20 crc kubenswrapper[4954]: I1209 18:36:20.702626 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_be42aebb-ed72-4e33-8e02-d8baace9c46d/loki-index-gateway/0.log" Dec 09 18:36:20 crc kubenswrapper[4954]: I1209 18:36:20.889315 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_392d7392-3bc6-42f1-91b5-56f00fc1d1c4/loki-ingester/0.log" Dec 09 18:36:20 crc kubenswrapper[4954]: I1209 18:36:20.936390 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-stgnv_f7dd2098-b496-49e7-9cdb-f87431dc135f/loki-querier/0.log" Dec 09 18:36:21 crc kubenswrapper[4954]: I1209 18:36:21.089076 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-gzndd_73361aee-2741-4328-b840-bcc3619f8cc0/loki-query-frontend/0.log" Dec 09 18:36:21 crc kubenswrapper[4954]: E1209 18:36:21.122473 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:36:34 crc kubenswrapper[4954]: E1209 18:36:34.129572 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:36:35 crc kubenswrapper[4954]: E1209 18:36:35.121851 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:36:35 crc kubenswrapper[4954]: I1209 18:36:35.972966 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rtqnp_fa6649f3-3933-4395-a345-f1a61a8616b1/kube-rbac-proxy/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.073088 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rtqnp_fa6649f3-3933-4395-a345-f1a61a8616b1/controller/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.177320 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-frr-files/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.369221 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-frr-files/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.390728 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-reloader/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.409243 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-metrics/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.436502 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-reloader/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.697273 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-metrics/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.707699 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-frr-files/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.708173 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-metrics/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.751369 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-reloader/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.904818 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-frr-files/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.934267 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-reloader/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.956137 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/controller/0.log" Dec 09 18:36:36 crc kubenswrapper[4954]: I1209 18:36:36.962946 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/cp-metrics/0.log" Dec 09 18:36:37 crc kubenswrapper[4954]: I1209 18:36:37.159800 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/kube-rbac-proxy/0.log" Dec 09 18:36:37 crc kubenswrapper[4954]: I1209 18:36:37.162900 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/frr-metrics/0.log" Dec 09 18:36:37 crc kubenswrapper[4954]: I1209 18:36:37.169361 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/kube-rbac-proxy-frr/0.log" Dec 09 18:36:37 crc kubenswrapper[4954]: I1209 18:36:37.326898 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/reloader/0.log" Dec 09 18:36:37 crc kubenswrapper[4954]: I1209 18:36:37.509138 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-f85kg_4b232ff4-1dc9-4a9a-a67a-ae42c47c99b6/frr-k8s-webhook-server/0.log" Dec 09 18:36:37 crc kubenswrapper[4954]: I1209 18:36:37.754670 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7fb7cc7845-b4jd8_4d80ef44-a45a-4e4f-b61f-d2b30d9c1fc2/manager/0.log" Dec 09 18:36:37 crc kubenswrapper[4954]: I1209 18:36:37.808620 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6b9d7ddc4d-l8z78_f8aed5f9-f76a-4a19-92ae-0b11a2a3fee7/webhook-server/0.log" Dec 09 18:36:37 crc kubenswrapper[4954]: I1209 18:36:37.992174 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wl7gt_87be1fd5-5517-4e21-b5df-807a44761f4e/kube-rbac-proxy/0.log" Dec 09 18:36:38 crc kubenswrapper[4954]: I1209 18:36:38.529277 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wl7gt_87be1fd5-5517-4e21-b5df-807a44761f4e/speaker/0.log" Dec 09 18:36:38 crc kubenswrapper[4954]: I1209 18:36:38.795820 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t56bj_b6f3c78e-58d6-4bba-a14e-85e3acd23e24/frr/0.log" Dec 09 18:36:43 crc kubenswrapper[4954]: I1209 18:36:43.755472 4954 patch_prober.go:28] interesting pod/machine-config-daemon-clkc5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 18:36:43 crc kubenswrapper[4954]: I1209 18:36:43.756205 4954 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 18:36:43 crc kubenswrapper[4954]: I1209 18:36:43.756268 4954 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" Dec 09 18:36:43 crc kubenswrapper[4954]: I1209 18:36:43.757242 4954 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2"} pod="openshift-machine-config-operator/machine-config-daemon-clkc5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 18:36:43 crc kubenswrapper[4954]: I1209 18:36:43.757306 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerName="machine-config-daemon" containerID="cri-o://fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" gracePeriod=600 Dec 09 18:36:43 crc kubenswrapper[4954]: E1209 18:36:43.880797 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:36:43 crc kubenswrapper[4954]: I1209 18:36:43.902389 4954 generic.go:334] "Generic (PLEG): container finished" podID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" exitCode=0 Dec 09 18:36:43 crc kubenswrapper[4954]: I1209 18:36:43.902429 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerDied","Data":"fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2"} Dec 09 18:36:43 crc kubenswrapper[4954]: I1209 18:36:43.902460 4954 scope.go:117] "RemoveContainer" containerID="afd4d72dcad9520c4cb369ae68cbac6019ac340837b03fa2d274871625d5a7a8" Dec 09 18:36:43 crc kubenswrapper[4954]: I1209 18:36:43.903197 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:36:43 crc kubenswrapper[4954]: E1209 18:36:43.903515 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:36:46 crc kubenswrapper[4954]: E1209 18:36:46.122797 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:36:47 crc kubenswrapper[4954]: E1209 18:36:47.122261 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:36:50 crc kubenswrapper[4954]: I1209 18:36:50.236183 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv_a457e4b5-7bbb-480a-ae97-7cdfc2822dec/util/0.log" Dec 09 18:36:50 crc kubenswrapper[4954]: I1209 18:36:50.368196 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv_a457e4b5-7bbb-480a-ae97-7cdfc2822dec/util/0.log" Dec 09 18:36:50 crc kubenswrapper[4954]: I1209 18:36:50.381377 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv_a457e4b5-7bbb-480a-ae97-7cdfc2822dec/pull/0.log" Dec 09 18:36:50 crc kubenswrapper[4954]: I1209 18:36:50.441471 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv_a457e4b5-7bbb-480a-ae97-7cdfc2822dec/pull/0.log" Dec 09 18:36:50 crc kubenswrapper[4954]: I1209 18:36:50.587494 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv_a457e4b5-7bbb-480a-ae97-7cdfc2822dec/util/0.log" Dec 09 18:36:50 crc kubenswrapper[4954]: I1209 18:36:50.612609 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv_a457e4b5-7bbb-480a-ae97-7cdfc2822dec/extract/0.log" Dec 09 18:36:50 crc kubenswrapper[4954]: I1209 18:36:50.623641 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ffmjlv_a457e4b5-7bbb-480a-ae97-7cdfc2822dec/pull/0.log" Dec 09 18:36:50 crc kubenswrapper[4954]: I1209 18:36:50.734005 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m_97ccf7f7-c707-4236-bb60-ac38713c8793/util/0.log" Dec 09 18:36:50 crc kubenswrapper[4954]: I1209 18:36:50.978436 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m_97ccf7f7-c707-4236-bb60-ac38713c8793/pull/0.log" Dec 09 18:36:50 crc kubenswrapper[4954]: I1209 18:36:50.984510 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m_97ccf7f7-c707-4236-bb60-ac38713c8793/util/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.001818 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m_97ccf7f7-c707-4236-bb60-ac38713c8793/pull/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.192560 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m_97ccf7f7-c707-4236-bb60-ac38713c8793/util/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.195033 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m_97ccf7f7-c707-4236-bb60-ac38713c8793/extract/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.202040 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83kp29m_97ccf7f7-c707-4236-bb60-ac38713c8793/pull/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.350013 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cbsp7_0055c556-4fda-424a-b590-5a6ca19d8314/extract-utilities/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.524583 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cbsp7_0055c556-4fda-424a-b590-5a6ca19d8314/extract-utilities/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.559315 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cbsp7_0055c556-4fda-424a-b590-5a6ca19d8314/extract-content/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.559731 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cbsp7_0055c556-4fda-424a-b590-5a6ca19d8314/extract-content/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.716329 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cbsp7_0055c556-4fda-424a-b590-5a6ca19d8314/extract-utilities/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.726792 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cbsp7_0055c556-4fda-424a-b590-5a6ca19d8314/extract-content/0.log" Dec 09 18:36:51 crc kubenswrapper[4954]: I1209 18:36:51.924904 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j4rwh_0ffead20-0858-4e11-8d58-2c85c1267cd6/extract-utilities/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.137292 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j4rwh_0ffead20-0858-4e11-8d58-2c85c1267cd6/extract-content/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.150705 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j4rwh_0ffead20-0858-4e11-8d58-2c85c1267cd6/extract-utilities/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.165075 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cbsp7_0055c556-4fda-424a-b590-5a6ca19d8314/registry-server/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.175391 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j4rwh_0ffead20-0858-4e11-8d58-2c85c1267cd6/extract-content/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.390242 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j4rwh_0ffead20-0858-4e11-8d58-2c85c1267cd6/extract-utilities/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.402169 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j4rwh_0ffead20-0858-4e11-8d58-2c85c1267cd6/extract-content/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.727675 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-krpqr_315ebcec-7a31-4e51-9f9b-46f715643fe5/extract-utilities/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.730934 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4svl9_2fd60f62-8fed-4f13-9da7-926b99b7d4aa/marketplace-operator/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.792331 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j4rwh_0ffead20-0858-4e11-8d58-2c85c1267cd6/registry-server/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.924275 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-krpqr_315ebcec-7a31-4e51-9f9b-46f715643fe5/extract-content/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.945242 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-krpqr_315ebcec-7a31-4e51-9f9b-46f715643fe5/extract-utilities/0.log" Dec 09 18:36:52 crc kubenswrapper[4954]: I1209 18:36:52.950642 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-krpqr_315ebcec-7a31-4e51-9f9b-46f715643fe5/extract-content/0.log" Dec 09 18:36:53 crc kubenswrapper[4954]: I1209 18:36:53.108968 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-krpqr_315ebcec-7a31-4e51-9f9b-46f715643fe5/extract-content/0.log" Dec 09 18:36:53 crc kubenswrapper[4954]: I1209 18:36:53.112920 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-krpqr_315ebcec-7a31-4e51-9f9b-46f715643fe5/extract-utilities/0.log" Dec 09 18:36:53 crc kubenswrapper[4954]: I1209 18:36:53.359542 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-krpqr_315ebcec-7a31-4e51-9f9b-46f715643fe5/registry-server/0.log" Dec 09 18:36:53 crc kubenswrapper[4954]: I1209 18:36:53.371303 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gg7lf_da760521-0631-4573-a06c-dc5bbf6dddc7/extract-utilities/0.log" Dec 09 18:36:53 crc kubenswrapper[4954]: I1209 18:36:53.517875 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gg7lf_da760521-0631-4573-a06c-dc5bbf6dddc7/extract-utilities/0.log" Dec 09 18:36:53 crc kubenswrapper[4954]: I1209 18:36:53.528197 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gg7lf_da760521-0631-4573-a06c-dc5bbf6dddc7/extract-content/0.log" Dec 09 18:36:53 crc kubenswrapper[4954]: I1209 18:36:53.547756 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gg7lf_da760521-0631-4573-a06c-dc5bbf6dddc7/extract-content/0.log" Dec 09 18:36:53 crc kubenswrapper[4954]: I1209 18:36:53.703421 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gg7lf_da760521-0631-4573-a06c-dc5bbf6dddc7/extract-content/0.log" Dec 09 18:36:53 crc kubenswrapper[4954]: I1209 18:36:53.739166 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gg7lf_da760521-0631-4573-a06c-dc5bbf6dddc7/extract-utilities/0.log" Dec 09 18:36:54 crc kubenswrapper[4954]: I1209 18:36:54.498221 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gg7lf_da760521-0631-4573-a06c-dc5bbf6dddc7/registry-server/0.log" Dec 09 18:36:57 crc kubenswrapper[4954]: I1209 18:36:57.120008 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:36:57 crc kubenswrapper[4954]: E1209 18:36:57.121631 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:36:58 crc kubenswrapper[4954]: E1209 18:36:58.123162 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:37:01 crc kubenswrapper[4954]: E1209 18:37:01.122897 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:37:05 crc kubenswrapper[4954]: I1209 18:37:05.022553 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-gtwjn_0e70cc7a-1b64-47c6-b0f1-4da0e77c9870/prometheus-operator/0.log" Dec 09 18:37:05 crc kubenswrapper[4954]: I1209 18:37:05.239719 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-648fd67fbd-cf59m_ab2760a1-3695-4556-bff0-d61d9da9d78b/prometheus-operator-admission-webhook/0.log" Dec 09 18:37:05 crc kubenswrapper[4954]: I1209 18:37:05.244378 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-648fd67fbd-7cqrn_71312e1a-0fc4-417a-92e3-080c5651ff47/prometheus-operator-admission-webhook/0.log" Dec 09 18:37:05 crc kubenswrapper[4954]: I1209 18:37:05.424514 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-87ttl_cdd1ba1a-7a70-48e0-914c-2bbe8c8135d6/operator/0.log" Dec 09 18:37:05 crc kubenswrapper[4954]: I1209 18:37:05.459293 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-mmts5_300e9948-95ee-44e4-b816-16664d86e2e2/observability-ui-dashboards/0.log" Dec 09 18:37:05 crc kubenswrapper[4954]: I1209 18:37:05.570325 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-2l8fr_2466d04c-561e-4c13-9939-9194d26ac50f/perses-operator/0.log" Dec 09 18:37:09 crc kubenswrapper[4954]: I1209 18:37:09.120613 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:37:09 crc kubenswrapper[4954]: E1209 18:37:09.122036 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:37:09 crc kubenswrapper[4954]: E1209 18:37:09.122869 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:37:13 crc kubenswrapper[4954]: E1209 18:37:13.123106 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.264408 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7hghf"] Dec 09 18:37:14 crc kubenswrapper[4954]: E1209 18:37:14.265982 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerName="extract-content" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.266091 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerName="extract-content" Dec 09 18:37:14 crc kubenswrapper[4954]: E1209 18:37:14.266744 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c15dbc-04f7-43a4-8aaf-2adda321c720" containerName="container-00" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.266777 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c15dbc-04f7-43a4-8aaf-2adda321c720" containerName="container-00" Dec 09 18:37:14 crc kubenswrapper[4954]: E1209 18:37:14.266813 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerName="extract-utilities" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.266820 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerName="extract-utilities" Dec 09 18:37:14 crc kubenswrapper[4954]: E1209 18:37:14.266849 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerName="registry-server" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.266855 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerName="registry-server" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.267275 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="b684cdd8-1c8e-4a1f-9b0c-98678ffe4168" containerName="registry-server" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.267295 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c15dbc-04f7-43a4-8aaf-2adda321c720" containerName="container-00" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.268988 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.279687 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7hghf"] Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.325985 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-utilities\") pod \"redhat-operators-7hghf\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.326082 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-catalog-content\") pod \"redhat-operators-7hghf\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.326220 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87p4r\" (UniqueName: \"kubernetes.io/projected/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-kube-api-access-87p4r\") pod \"redhat-operators-7hghf\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.427803 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-catalog-content\") pod \"redhat-operators-7hghf\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.427988 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87p4r\" (UniqueName: \"kubernetes.io/projected/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-kube-api-access-87p4r\") pod \"redhat-operators-7hghf\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.428076 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-utilities\") pod \"redhat-operators-7hghf\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.428368 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-catalog-content\") pod \"redhat-operators-7hghf\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.428448 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-utilities\") pod \"redhat-operators-7hghf\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.460509 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87p4r\" (UniqueName: \"kubernetes.io/projected/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-kube-api-access-87p4r\") pod \"redhat-operators-7hghf\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:14 crc kubenswrapper[4954]: I1209 18:37:14.597441 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:15 crc kubenswrapper[4954]: I1209 18:37:15.127785 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7hghf"] Dec 09 18:37:15 crc kubenswrapper[4954]: I1209 18:37:15.213418 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hghf" event={"ID":"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d","Type":"ContainerStarted","Data":"5f8bbcfa156c6e72f8e03bab1a2300354991e72aa60c9a5ea7de86770fa5bfc9"} Dec 09 18:37:16 crc kubenswrapper[4954]: I1209 18:37:16.240888 4954 generic.go:334] "Generic (PLEG): container finished" podID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerID="2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1" exitCode=0 Dec 09 18:37:16 crc kubenswrapper[4954]: I1209 18:37:16.240953 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hghf" event={"ID":"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d","Type":"ContainerDied","Data":"2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1"} Dec 09 18:37:17 crc kubenswrapper[4954]: I1209 18:37:17.254368 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hghf" event={"ID":"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d","Type":"ContainerStarted","Data":"b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3"} Dec 09 18:37:17 crc kubenswrapper[4954]: I1209 18:37:17.906994 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5f76df48c6-fvg95_73f81f6d-15c4-4cc8-95be-59c4db2139ea/kube-rbac-proxy/0.log" Dec 09 18:37:17 crc kubenswrapper[4954]: I1209 18:37:17.920260 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5f76df48c6-fvg95_73f81f6d-15c4-4cc8-95be-59c4db2139ea/manager/0.log" Dec 09 18:37:21 crc kubenswrapper[4954]: I1209 18:37:21.121801 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:37:21 crc kubenswrapper[4954]: E1209 18:37:21.122676 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:37:21 crc kubenswrapper[4954]: E1209 18:37:21.126547 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:37:21 crc kubenswrapper[4954]: I1209 18:37:21.303723 4954 generic.go:334] "Generic (PLEG): container finished" podID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerID="b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3" exitCode=0 Dec 09 18:37:21 crc kubenswrapper[4954]: I1209 18:37:21.303793 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hghf" event={"ID":"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d","Type":"ContainerDied","Data":"b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3"} Dec 09 18:37:22 crc kubenswrapper[4954]: I1209 18:37:22.318291 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hghf" event={"ID":"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d","Type":"ContainerStarted","Data":"3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61"} Dec 09 18:37:22 crc kubenswrapper[4954]: I1209 18:37:22.341915 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7hghf" podStartSLOduration=2.77617244 podStartE2EDuration="8.341886466s" podCreationTimestamp="2025-12-09 18:37:14 +0000 UTC" firstStartedPulling="2025-12-09 18:37:16.245379674 +0000 UTC m=+6032.633553494" lastFinishedPulling="2025-12-09 18:37:21.8110937 +0000 UTC m=+6038.199267520" observedRunningTime="2025-12-09 18:37:22.336769386 +0000 UTC m=+6038.724943216" watchObservedRunningTime="2025-12-09 18:37:22.341886466 +0000 UTC m=+6038.730060286" Dec 09 18:37:24 crc kubenswrapper[4954]: I1209 18:37:24.597796 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:24 crc kubenswrapper[4954]: I1209 18:37:24.599340 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:25 crc kubenswrapper[4954]: I1209 18:37:25.678071 4954 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7hghf" podUID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerName="registry-server" probeResult="failure" output=< Dec 09 18:37:25 crc kubenswrapper[4954]: timeout: failed to connect service ":50051" within 1s Dec 09 18:37:25 crc kubenswrapper[4954]: > Dec 09 18:37:28 crc kubenswrapper[4954]: E1209 18:37:28.123290 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:37:34 crc kubenswrapper[4954]: I1209 18:37:34.661508 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:34 crc kubenswrapper[4954]: I1209 18:37:34.711934 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:34 crc kubenswrapper[4954]: I1209 18:37:34.899726 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7hghf"] Dec 09 18:37:35 crc kubenswrapper[4954]: E1209 18:37:35.121949 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:37:36 crc kubenswrapper[4954]: I1209 18:37:36.120184 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:37:36 crc kubenswrapper[4954]: E1209 18:37:36.121292 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:37:36 crc kubenswrapper[4954]: E1209 18:37:36.297808 4954 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.174:52866->38.102.83.174:40929: write tcp 38.102.83.174:52866->38.102.83.174:40929: write: broken pipe Dec 09 18:37:36 crc kubenswrapper[4954]: I1209 18:37:36.454205 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7hghf" podUID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerName="registry-server" containerID="cri-o://3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61" gracePeriod=2 Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.006258 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.108728 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-utilities\") pod \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.109002 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87p4r\" (UniqueName: \"kubernetes.io/projected/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-kube-api-access-87p4r\") pod \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.109061 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-catalog-content\") pod \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\" (UID: \"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d\") " Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.110078 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-utilities" (OuterVolumeSpecName: "utilities") pod "56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" (UID: "56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.116787 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-kube-api-access-87p4r" (OuterVolumeSpecName: "kube-api-access-87p4r") pod "56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" (UID: "56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d"). InnerVolumeSpecName "kube-api-access-87p4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.211299 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87p4r\" (UniqueName: \"kubernetes.io/projected/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-kube-api-access-87p4r\") on node \"crc\" DevicePath \"\"" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.211333 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.254173 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" (UID: "56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.313080 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.470227 4954 generic.go:334] "Generic (PLEG): container finished" podID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerID="3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61" exitCode=0 Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.470269 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hghf" event={"ID":"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d","Type":"ContainerDied","Data":"3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61"} Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.470303 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hghf" event={"ID":"56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d","Type":"ContainerDied","Data":"5f8bbcfa156c6e72f8e03bab1a2300354991e72aa60c9a5ea7de86770fa5bfc9"} Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.470304 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7hghf" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.470328 4954 scope.go:117] "RemoveContainer" containerID="3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.503355 4954 scope.go:117] "RemoveContainer" containerID="b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.525864 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7hghf"] Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.537034 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7hghf"] Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.543254 4954 scope.go:117] "RemoveContainer" containerID="2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.585853 4954 scope.go:117] "RemoveContainer" containerID="3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61" Dec 09 18:37:37 crc kubenswrapper[4954]: E1209 18:37:37.586341 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61\": container with ID starting with 3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61 not found: ID does not exist" containerID="3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.586373 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61"} err="failed to get container status \"3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61\": rpc error: code = NotFound desc = could not find container \"3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61\": container with ID starting with 3f615e333a1803ee96390a6df867df04fd33663f872de0224295f9ada6415e61 not found: ID does not exist" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.586400 4954 scope.go:117] "RemoveContainer" containerID="b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3" Dec 09 18:37:37 crc kubenswrapper[4954]: E1209 18:37:37.586972 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3\": container with ID starting with b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3 not found: ID does not exist" containerID="b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.587002 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3"} err="failed to get container status \"b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3\": rpc error: code = NotFound desc = could not find container \"b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3\": container with ID starting with b36da19ab3d9b3f298c254e1fde18958726c9db05f67d29687e8d6a3e6beedc3 not found: ID does not exist" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.587020 4954 scope.go:117] "RemoveContainer" containerID="2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1" Dec 09 18:37:37 crc kubenswrapper[4954]: E1209 18:37:37.587382 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1\": container with ID starting with 2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1 not found: ID does not exist" containerID="2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1" Dec 09 18:37:37 crc kubenswrapper[4954]: I1209 18:37:37.587427 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1"} err="failed to get container status \"2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1\": rpc error: code = NotFound desc = could not find container \"2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1\": container with ID starting with 2fbd575ac1597252c29890e080ef7d23b4aba3845b2d19f22dca9a94b0c712f1 not found: ID does not exist" Dec 09 18:37:38 crc kubenswrapper[4954]: I1209 18:37:38.138815 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" path="/var/lib/kubelet/pods/56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d/volumes" Dec 09 18:37:41 crc kubenswrapper[4954]: E1209 18:37:41.122822 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:37:48 crc kubenswrapper[4954]: I1209 18:37:48.122871 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:37:48 crc kubenswrapper[4954]: E1209 18:37:48.125083 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:37:49 crc kubenswrapper[4954]: E1209 18:37:49.123032 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:37:56 crc kubenswrapper[4954]: E1209 18:37:56.124678 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:38:00 crc kubenswrapper[4954]: I1209 18:38:00.121040 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:38:00 crc kubenswrapper[4954]: E1209 18:38:00.121860 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:38:01 crc kubenswrapper[4954]: E1209 18:38:01.122992 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:38:08 crc kubenswrapper[4954]: E1209 18:38:08.122240 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:38:11 crc kubenswrapper[4954]: I1209 18:38:11.121210 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:38:11 crc kubenswrapper[4954]: E1209 18:38:11.121966 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:38:13 crc kubenswrapper[4954]: E1209 18:38:13.125143 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:38:21 crc kubenswrapper[4954]: E1209 18:38:21.123493 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:38:24 crc kubenswrapper[4954]: I1209 18:38:24.128704 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:38:24 crc kubenswrapper[4954]: E1209 18:38:24.129318 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:38:27 crc kubenswrapper[4954]: E1209 18:38:27.151708 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:38:35 crc kubenswrapper[4954]: E1209 18:38:35.122805 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:38:36 crc kubenswrapper[4954]: I1209 18:38:36.121212 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:38:36 crc kubenswrapper[4954]: E1209 18:38:36.121941 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:38:41 crc kubenswrapper[4954]: I1209 18:38:41.123921 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 18:38:41 crc kubenswrapper[4954]: E1209 18:38:41.257028 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:38:41 crc kubenswrapper[4954]: E1209 18:38:41.257113 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:38:41 crc kubenswrapper[4954]: E1209 18:38:41.257366 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:38:41 crc kubenswrapper[4954]: E1209 18:38:41.258660 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:38:46 crc kubenswrapper[4954]: E1209 18:38:46.123752 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:38:51 crc kubenswrapper[4954]: I1209 18:38:51.120347 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:38:51 crc kubenswrapper[4954]: E1209 18:38:51.121203 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:38:54 crc kubenswrapper[4954]: E1209 18:38:54.140306 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:38:54 crc kubenswrapper[4954]: I1209 18:38:54.221972 4954 generic.go:334] "Generic (PLEG): container finished" podID="9b11b7cc-7b44-4800-95b1-37f4cd8a6146" containerID="146f170dd8360de9b5092b06ed52724ce53e6a215cfec261d8d8471a61271a20" exitCode=0 Dec 09 18:38:54 crc kubenswrapper[4954]: I1209 18:38:54.222014 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hlk28/must-gather-b4g8k" event={"ID":"9b11b7cc-7b44-4800-95b1-37f4cd8a6146","Type":"ContainerDied","Data":"146f170dd8360de9b5092b06ed52724ce53e6a215cfec261d8d8471a61271a20"} Dec 09 18:38:54 crc kubenswrapper[4954]: I1209 18:38:54.222807 4954 scope.go:117] "RemoveContainer" containerID="146f170dd8360de9b5092b06ed52724ce53e6a215cfec261d8d8471a61271a20" Dec 09 18:38:54 crc kubenswrapper[4954]: I1209 18:38:54.321181 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hlk28_must-gather-b4g8k_9b11b7cc-7b44-4800-95b1-37f4cd8a6146/gather/0.log" Dec 09 18:38:59 crc kubenswrapper[4954]: E1209 18:38:59.251925 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:38:59 crc kubenswrapper[4954]: E1209 18:38:59.252515 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 09 18:38:59 crc kubenswrapper[4954]: E1209 18:38:59.252726 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz7bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-b2pg9_openstack(8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:38:59 crc kubenswrapper[4954]: E1209 18:38:59.253851 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-heat-engine: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.011969 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hlk28/must-gather-b4g8k"] Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.012540 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-hlk28/must-gather-b4g8k" podUID="9b11b7cc-7b44-4800-95b1-37f4cd8a6146" containerName="copy" containerID="cri-o://3c165ce3af65b7ee339bdb0106e348eba8c01f6f953634bbaffcde8d99b22101" gracePeriod=2 Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.023493 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hlk28/must-gather-b4g8k"] Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.306345 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hlk28_must-gather-b4g8k_9b11b7cc-7b44-4800-95b1-37f4cd8a6146/copy/0.log" Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.306790 4954 generic.go:334] "Generic (PLEG): container finished" podID="9b11b7cc-7b44-4800-95b1-37f4cd8a6146" containerID="3c165ce3af65b7ee339bdb0106e348eba8c01f6f953634bbaffcde8d99b22101" exitCode=143 Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.469703 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hlk28_must-gather-b4g8k_9b11b7cc-7b44-4800-95b1-37f4cd8a6146/copy/0.log" Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.472148 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/must-gather-b4g8k" Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.583844 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gc5c\" (UniqueName: \"kubernetes.io/projected/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-kube-api-access-4gc5c\") pod \"9b11b7cc-7b44-4800-95b1-37f4cd8a6146\" (UID: \"9b11b7cc-7b44-4800-95b1-37f4cd8a6146\") " Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.583914 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-must-gather-output\") pod \"9b11b7cc-7b44-4800-95b1-37f4cd8a6146\" (UID: \"9b11b7cc-7b44-4800-95b1-37f4cd8a6146\") " Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.590061 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-kube-api-access-4gc5c" (OuterVolumeSpecName: "kube-api-access-4gc5c") pod "9b11b7cc-7b44-4800-95b1-37f4cd8a6146" (UID: "9b11b7cc-7b44-4800-95b1-37f4cd8a6146"). InnerVolumeSpecName "kube-api-access-4gc5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.686420 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gc5c\" (UniqueName: \"kubernetes.io/projected/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-kube-api-access-4gc5c\") on node \"crc\" DevicePath \"\"" Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.761432 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "9b11b7cc-7b44-4800-95b1-37f4cd8a6146" (UID: "9b11b7cc-7b44-4800-95b1-37f4cd8a6146"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:39:02 crc kubenswrapper[4954]: I1209 18:39:02.789201 4954 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9b11b7cc-7b44-4800-95b1-37f4cd8a6146-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 09 18:39:03 crc kubenswrapper[4954]: I1209 18:39:03.318919 4954 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hlk28_must-gather-b4g8k_9b11b7cc-7b44-4800-95b1-37f4cd8a6146/copy/0.log" Dec 09 18:39:03 crc kubenswrapper[4954]: I1209 18:39:03.320319 4954 scope.go:117] "RemoveContainer" containerID="3c165ce3af65b7ee339bdb0106e348eba8c01f6f953634bbaffcde8d99b22101" Dec 09 18:39:03 crc kubenswrapper[4954]: I1209 18:39:03.320548 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hlk28/must-gather-b4g8k" Dec 09 18:39:03 crc kubenswrapper[4954]: I1209 18:39:03.343334 4954 scope.go:117] "RemoveContainer" containerID="146f170dd8360de9b5092b06ed52724ce53e6a215cfec261d8d8471a61271a20" Dec 09 18:39:04 crc kubenswrapper[4954]: I1209 18:39:04.132506 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b11b7cc-7b44-4800-95b1-37f4cd8a6146" path="/var/lib/kubelet/pods/9b11b7cc-7b44-4800-95b1-37f4cd8a6146/volumes" Dec 09 18:39:06 crc kubenswrapper[4954]: I1209 18:39:06.120414 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:39:06 crc kubenswrapper[4954]: E1209 18:39:06.121002 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:39:09 crc kubenswrapper[4954]: E1209 18:39:09.122976 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:39:12 crc kubenswrapper[4954]: E1209 18:39:12.128138 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:39:20 crc kubenswrapper[4954]: I1209 18:39:20.120925 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:39:20 crc kubenswrapper[4954]: E1209 18:39:20.121962 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:39:21 crc kubenswrapper[4954]: E1209 18:39:21.125192 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:39:25 crc kubenswrapper[4954]: E1209 18:39:25.133850 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:39:32 crc kubenswrapper[4954]: E1209 18:39:32.123441 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:39:35 crc kubenswrapper[4954]: I1209 18:39:35.121579 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:39:35 crc kubenswrapper[4954]: E1209 18:39:35.122217 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:39:38 crc kubenswrapper[4954]: E1209 18:39:38.122914 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.379480 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fctzq"] Dec 09 18:39:41 crc kubenswrapper[4954]: E1209 18:39:41.380492 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerName="registry-server" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.380509 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerName="registry-server" Dec 09 18:39:41 crc kubenswrapper[4954]: E1209 18:39:41.380522 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b11b7cc-7b44-4800-95b1-37f4cd8a6146" containerName="copy" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.380527 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b11b7cc-7b44-4800-95b1-37f4cd8a6146" containerName="copy" Dec 09 18:39:41 crc kubenswrapper[4954]: E1209 18:39:41.380539 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerName="extract-utilities" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.380545 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerName="extract-utilities" Dec 09 18:39:41 crc kubenswrapper[4954]: E1209 18:39:41.380561 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b11b7cc-7b44-4800-95b1-37f4cd8a6146" containerName="gather" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.380567 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b11b7cc-7b44-4800-95b1-37f4cd8a6146" containerName="gather" Dec 09 18:39:41 crc kubenswrapper[4954]: E1209 18:39:41.380646 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerName="extract-content" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.380654 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerName="extract-content" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.380871 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b11b7cc-7b44-4800-95b1-37f4cd8a6146" containerName="gather" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.380882 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b11b7cc-7b44-4800-95b1-37f4cd8a6146" containerName="copy" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.380898 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="56a4a1e6-3067-46f6-9cfd-c6a9d7b4b94d" containerName="registry-server" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.382696 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.407257 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fctzq"] Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.491614 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-catalog-content\") pod \"redhat-marketplace-fctzq\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.491737 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9c54\" (UniqueName: \"kubernetes.io/projected/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-kube-api-access-x9c54\") pod \"redhat-marketplace-fctzq\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.492195 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-utilities\") pod \"redhat-marketplace-fctzq\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.594738 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9c54\" (UniqueName: \"kubernetes.io/projected/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-kube-api-access-x9c54\") pod \"redhat-marketplace-fctzq\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.594932 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-utilities\") pod \"redhat-marketplace-fctzq\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.595000 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-catalog-content\") pod \"redhat-marketplace-fctzq\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.595630 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-utilities\") pod \"redhat-marketplace-fctzq\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.595651 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-catalog-content\") pod \"redhat-marketplace-fctzq\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.621464 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9c54\" (UniqueName: \"kubernetes.io/projected/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-kube-api-access-x9c54\") pod \"redhat-marketplace-fctzq\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:41 crc kubenswrapper[4954]: I1209 18:39:41.701767 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:42 crc kubenswrapper[4954]: I1209 18:39:42.233521 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fctzq"] Dec 09 18:39:42 crc kubenswrapper[4954]: I1209 18:39:42.825852 4954 generic.go:334] "Generic (PLEG): container finished" podID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerID="3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d" exitCode=0 Dec 09 18:39:42 crc kubenswrapper[4954]: I1209 18:39:42.825901 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fctzq" event={"ID":"d7ceb089-e860-4856-bb4f-e3e7e5b242d0","Type":"ContainerDied","Data":"3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d"} Dec 09 18:39:42 crc kubenswrapper[4954]: I1209 18:39:42.825928 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fctzq" event={"ID":"d7ceb089-e860-4856-bb4f-e3e7e5b242d0","Type":"ContainerStarted","Data":"10d43336f787a9cf563e7589f4927c438274c82f41e7144b5be51d1cca67ef90"} Dec 09 18:39:44 crc kubenswrapper[4954]: I1209 18:39:44.847725 4954 generic.go:334] "Generic (PLEG): container finished" podID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerID="7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30" exitCode=0 Dec 09 18:39:44 crc kubenswrapper[4954]: I1209 18:39:44.847807 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fctzq" event={"ID":"d7ceb089-e860-4856-bb4f-e3e7e5b242d0","Type":"ContainerDied","Data":"7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30"} Dec 09 18:39:45 crc kubenswrapper[4954]: E1209 18:39:45.121996 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:39:45 crc kubenswrapper[4954]: I1209 18:39:45.859291 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fctzq" event={"ID":"d7ceb089-e860-4856-bb4f-e3e7e5b242d0","Type":"ContainerStarted","Data":"7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49"} Dec 09 18:39:45 crc kubenswrapper[4954]: I1209 18:39:45.884556 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fctzq" podStartSLOduration=2.298488374 podStartE2EDuration="4.884534025s" podCreationTimestamp="2025-12-09 18:39:41 +0000 UTC" firstStartedPulling="2025-12-09 18:39:42.827441531 +0000 UTC m=+6179.215615351" lastFinishedPulling="2025-12-09 18:39:45.413487182 +0000 UTC m=+6181.801661002" observedRunningTime="2025-12-09 18:39:45.880252891 +0000 UTC m=+6182.268426711" watchObservedRunningTime="2025-12-09 18:39:45.884534025 +0000 UTC m=+6182.272707845" Dec 09 18:39:47 crc kubenswrapper[4954]: I1209 18:39:47.120491 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:39:47 crc kubenswrapper[4954]: E1209 18:39:47.121118 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:39:51 crc kubenswrapper[4954]: E1209 18:39:51.122641 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:39:51 crc kubenswrapper[4954]: I1209 18:39:51.703272 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:51 crc kubenswrapper[4954]: I1209 18:39:51.703542 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:51 crc kubenswrapper[4954]: I1209 18:39:51.764448 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:51 crc kubenswrapper[4954]: I1209 18:39:51.975947 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:52 crc kubenswrapper[4954]: I1209 18:39:52.033646 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fctzq"] Dec 09 18:39:53 crc kubenswrapper[4954]: I1209 18:39:53.942612 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fctzq" podUID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerName="registry-server" containerID="cri-o://7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49" gracePeriod=2 Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.453160 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.602116 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-catalog-content\") pod \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.602208 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9c54\" (UniqueName: \"kubernetes.io/projected/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-kube-api-access-x9c54\") pod \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.602481 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-utilities\") pod \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\" (UID: \"d7ceb089-e860-4856-bb4f-e3e7e5b242d0\") " Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.603417 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-utilities" (OuterVolumeSpecName: "utilities") pod "d7ceb089-e860-4856-bb4f-e3e7e5b242d0" (UID: "d7ceb089-e860-4856-bb4f-e3e7e5b242d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.604200 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.608171 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-kube-api-access-x9c54" (OuterVolumeSpecName: "kube-api-access-x9c54") pod "d7ceb089-e860-4856-bb4f-e3e7e5b242d0" (UID: "d7ceb089-e860-4856-bb4f-e3e7e5b242d0"). InnerVolumeSpecName "kube-api-access-x9c54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.622628 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7ceb089-e860-4856-bb4f-e3e7e5b242d0" (UID: "d7ceb089-e860-4856-bb4f-e3e7e5b242d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.706579 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.706651 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9c54\" (UniqueName: \"kubernetes.io/projected/d7ceb089-e860-4856-bb4f-e3e7e5b242d0-kube-api-access-x9c54\") on node \"crc\" DevicePath \"\"" Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.956088 4954 generic.go:334] "Generic (PLEG): container finished" podID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerID="7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49" exitCode=0 Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.956144 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fctzq" event={"ID":"d7ceb089-e860-4856-bb4f-e3e7e5b242d0","Type":"ContainerDied","Data":"7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49"} Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.956177 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fctzq" event={"ID":"d7ceb089-e860-4856-bb4f-e3e7e5b242d0","Type":"ContainerDied","Data":"10d43336f787a9cf563e7589f4927c438274c82f41e7144b5be51d1cca67ef90"} Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.956195 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fctzq" Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.956201 4954 scope.go:117] "RemoveContainer" containerID="7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49" Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.980440 4954 scope.go:117] "RemoveContainer" containerID="7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30" Dec 09 18:39:54 crc kubenswrapper[4954]: I1209 18:39:54.995754 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fctzq"] Dec 09 18:39:55 crc kubenswrapper[4954]: I1209 18:39:55.007831 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fctzq"] Dec 09 18:39:55 crc kubenswrapper[4954]: I1209 18:39:55.024559 4954 scope.go:117] "RemoveContainer" containerID="3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d" Dec 09 18:39:55 crc kubenswrapper[4954]: I1209 18:39:55.059895 4954 scope.go:117] "RemoveContainer" containerID="7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49" Dec 09 18:39:55 crc kubenswrapper[4954]: E1209 18:39:55.061001 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49\": container with ID starting with 7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49 not found: ID does not exist" containerID="7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49" Dec 09 18:39:55 crc kubenswrapper[4954]: I1209 18:39:55.061058 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49"} err="failed to get container status \"7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49\": rpc error: code = NotFound desc = could not find container \"7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49\": container with ID starting with 7f69b46a5c03237d6c9bf7c9f0927aaf1d68ce6e893c8836545b9b7d835a4a49 not found: ID does not exist" Dec 09 18:39:55 crc kubenswrapper[4954]: I1209 18:39:55.061421 4954 scope.go:117] "RemoveContainer" containerID="7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30" Dec 09 18:39:55 crc kubenswrapper[4954]: E1209 18:39:55.061975 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30\": container with ID starting with 7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30 not found: ID does not exist" containerID="7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30" Dec 09 18:39:55 crc kubenswrapper[4954]: I1209 18:39:55.062087 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30"} err="failed to get container status \"7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30\": rpc error: code = NotFound desc = could not find container \"7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30\": container with ID starting with 7bdaef530f1c61d75bf186340f658ce645b639f513016f2dc0110d3aaaf4af30 not found: ID does not exist" Dec 09 18:39:55 crc kubenswrapper[4954]: I1209 18:39:55.062172 4954 scope.go:117] "RemoveContainer" containerID="3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d" Dec 09 18:39:55 crc kubenswrapper[4954]: E1209 18:39:55.062556 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d\": container with ID starting with 3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d not found: ID does not exist" containerID="3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d" Dec 09 18:39:55 crc kubenswrapper[4954]: I1209 18:39:55.062662 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d"} err="failed to get container status \"3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d\": rpc error: code = NotFound desc = could not find container \"3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d\": container with ID starting with 3f7e8ea9911a937b71a4b366b025a7d86e1d3de967634f187f9141d4a20aff2d not found: ID does not exist" Dec 09 18:39:56 crc kubenswrapper[4954]: I1209 18:39:56.134229 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" path="/var/lib/kubelet/pods/d7ceb089-e860-4856-bb4f-e3e7e5b242d0/volumes" Dec 09 18:39:57 crc kubenswrapper[4954]: E1209 18:39:57.122521 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:39:59 crc kubenswrapper[4954]: I1209 18:39:59.120473 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:39:59 crc kubenswrapper[4954]: E1209 18:39:59.121175 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:40:04 crc kubenswrapper[4954]: E1209 18:40:04.130243 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:40:07 crc kubenswrapper[4954]: I1209 18:40:07.139790 4954 scope.go:117] "RemoveContainer" containerID="4cf66fa15faeeed18d41beaa8cacba33d8bd82f4795edc2b3e300897adcb3ee0" Dec 09 18:40:10 crc kubenswrapper[4954]: I1209 18:40:10.120535 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:40:10 crc kubenswrapper[4954]: E1209 18:40:10.121320 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:40:12 crc kubenswrapper[4954]: E1209 18:40:12.122536 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:40:17 crc kubenswrapper[4954]: E1209 18:40:17.122080 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:40:22 crc kubenswrapper[4954]: I1209 18:40:22.120727 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:40:22 crc kubenswrapper[4954]: E1209 18:40:22.121470 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:40:27 crc kubenswrapper[4954]: E1209 18:40:27.124741 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:40:28 crc kubenswrapper[4954]: E1209 18:40:28.122709 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:40:36 crc kubenswrapper[4954]: I1209 18:40:36.120809 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:40:36 crc kubenswrapper[4954]: E1209 18:40:36.121589 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:40:42 crc kubenswrapper[4954]: E1209 18:40:42.122583 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:40:43 crc kubenswrapper[4954]: E1209 18:40:43.121367 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:40:47 crc kubenswrapper[4954]: I1209 18:40:47.119922 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:40:47 crc kubenswrapper[4954]: E1209 18:40:47.120707 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:40:53 crc kubenswrapper[4954]: E1209 18:40:53.123862 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:40:57 crc kubenswrapper[4954]: E1209 18:40:57.122558 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:40:58 crc kubenswrapper[4954]: I1209 18:40:58.120733 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:40:58 crc kubenswrapper[4954]: E1209 18:40:58.121412 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:41:05 crc kubenswrapper[4954]: E1209 18:41:05.122305 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:41:10 crc kubenswrapper[4954]: E1209 18:41:10.121865 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:41:11 crc kubenswrapper[4954]: I1209 18:41:11.120512 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:41:11 crc kubenswrapper[4954]: E1209 18:41:11.120965 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:41:16 crc kubenswrapper[4954]: E1209 18:41:16.123873 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:41:23 crc kubenswrapper[4954]: I1209 18:41:23.120885 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:41:23 crc kubenswrapper[4954]: E1209 18:41:23.121755 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:41:24 crc kubenswrapper[4954]: E1209 18:41:24.131235 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:41:29 crc kubenswrapper[4954]: E1209 18:41:29.122724 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:41:35 crc kubenswrapper[4954]: I1209 18:41:35.120120 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:41:35 crc kubenswrapper[4954]: E1209 18:41:35.121077 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-clkc5_openshift-machine-config-operator(5a6f4be9-a12a-4562-96b8-fd85694aa29c)\"" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" podUID="5a6f4be9-a12a-4562-96b8-fd85694aa29c" Dec 09 18:41:37 crc kubenswrapper[4954]: E1209 18:41:37.122614 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:41:41 crc kubenswrapper[4954]: E1209 18:41:41.123982 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:41:49 crc kubenswrapper[4954]: E1209 18:41:49.122539 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:41:50 crc kubenswrapper[4954]: I1209 18:41:50.120458 4954 scope.go:117] "RemoveContainer" containerID="fb12379164d321024d0fbf26316bde4c97dab912e1528cac9b29efe976f7d4a2" Dec 09 18:41:51 crc kubenswrapper[4954]: I1209 18:41:51.177268 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-clkc5" event={"ID":"5a6f4be9-a12a-4562-96b8-fd85694aa29c","Type":"ContainerStarted","Data":"f0bb4219874ee053a1d064e7d7e9ca102b0f089a9c32560a6e71328611a28fd2"} Dec 09 18:41:52 crc kubenswrapper[4954]: E1209 18:41:52.123018 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:42:04 crc kubenswrapper[4954]: E1209 18:42:04.128895 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:42:06 crc kubenswrapper[4954]: E1209 18:42:06.123748 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:42:15 crc kubenswrapper[4954]: E1209 18:42:15.123460 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:42:19 crc kubenswrapper[4954]: E1209 18:42:19.123119 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:42:27 crc kubenswrapper[4954]: E1209 18:42:27.122445 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:42:30 crc kubenswrapper[4954]: E1209 18:42:30.123330 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:42:38 crc kubenswrapper[4954]: E1209 18:42:38.123118 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:42:45 crc kubenswrapper[4954]: E1209 18:42:45.123903 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:42:49 crc kubenswrapper[4954]: E1209 18:42:49.122982 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:42:59 crc kubenswrapper[4954]: E1209 18:42:59.122269 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:43:02 crc kubenswrapper[4954]: E1209 18:43:02.122361 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:43:10 crc kubenswrapper[4954]: E1209 18:43:10.123785 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:43:16 crc kubenswrapper[4954]: E1209 18:43:16.122340 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.052639 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m49dd"] Dec 09 18:43:17 crc kubenswrapper[4954]: E1209 18:43:17.053557 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerName="extract-utilities" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.053580 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerName="extract-utilities" Dec 09 18:43:17 crc kubenswrapper[4954]: E1209 18:43:17.053634 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerName="registry-server" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.053644 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerName="registry-server" Dec 09 18:43:17 crc kubenswrapper[4954]: E1209 18:43:17.053659 4954 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerName="extract-content" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.053667 4954 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerName="extract-content" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.053913 4954 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7ceb089-e860-4856-bb4f-e3e7e5b242d0" containerName="registry-server" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.055562 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.066535 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m49dd"] Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.240616 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-utilities\") pod \"community-operators-m49dd\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.240668 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57xvd\" (UniqueName: \"kubernetes.io/projected/362670c4-832d-4388-a0e2-1e75c6e22738-kube-api-access-57xvd\") pod \"community-operators-m49dd\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.240987 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-catalog-content\") pod \"community-operators-m49dd\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.343825 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-catalog-content\") pod \"community-operators-m49dd\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.344336 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-catalog-content\") pod \"community-operators-m49dd\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.344489 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-utilities\") pod \"community-operators-m49dd\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.344553 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57xvd\" (UniqueName: \"kubernetes.io/projected/362670c4-832d-4388-a0e2-1e75c6e22738-kube-api-access-57xvd\") pod \"community-operators-m49dd\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.344964 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-utilities\") pod \"community-operators-m49dd\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.374690 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57xvd\" (UniqueName: \"kubernetes.io/projected/362670c4-832d-4388-a0e2-1e75c6e22738-kube-api-access-57xvd\") pod \"community-operators-m49dd\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.388067 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:17 crc kubenswrapper[4954]: I1209 18:43:17.946523 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m49dd"] Dec 09 18:43:17 crc kubenswrapper[4954]: W1209 18:43:17.955274 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod362670c4_832d_4388_a0e2_1e75c6e22738.slice/crio-dd546196a234d32bf72ebe4354dfcfe289e628ea15adae0bb70f4e3eec1fa0ed WatchSource:0}: Error finding container dd546196a234d32bf72ebe4354dfcfe289e628ea15adae0bb70f4e3eec1fa0ed: Status 404 returned error can't find the container with id dd546196a234d32bf72ebe4354dfcfe289e628ea15adae0bb70f4e3eec1fa0ed Dec 09 18:43:18 crc kubenswrapper[4954]: I1209 18:43:18.027104 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m49dd" event={"ID":"362670c4-832d-4388-a0e2-1e75c6e22738","Type":"ContainerStarted","Data":"dd546196a234d32bf72ebe4354dfcfe289e628ea15adae0bb70f4e3eec1fa0ed"} Dec 09 18:43:19 crc kubenswrapper[4954]: I1209 18:43:19.049959 4954 generic.go:334] "Generic (PLEG): container finished" podID="362670c4-832d-4388-a0e2-1e75c6e22738" containerID="a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc" exitCode=0 Dec 09 18:43:19 crc kubenswrapper[4954]: I1209 18:43:19.050230 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m49dd" event={"ID":"362670c4-832d-4388-a0e2-1e75c6e22738","Type":"ContainerDied","Data":"a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc"} Dec 09 18:43:20 crc kubenswrapper[4954]: I1209 18:43:20.068559 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m49dd" event={"ID":"362670c4-832d-4388-a0e2-1e75c6e22738","Type":"ContainerStarted","Data":"36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7"} Dec 09 18:43:21 crc kubenswrapper[4954]: I1209 18:43:21.078548 4954 generic.go:334] "Generic (PLEG): container finished" podID="362670c4-832d-4388-a0e2-1e75c6e22738" containerID="36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7" exitCode=0 Dec 09 18:43:21 crc kubenswrapper[4954]: I1209 18:43:21.078646 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m49dd" event={"ID":"362670c4-832d-4388-a0e2-1e75c6e22738","Type":"ContainerDied","Data":"36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7"} Dec 09 18:43:22 crc kubenswrapper[4954]: I1209 18:43:22.090767 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m49dd" event={"ID":"362670c4-832d-4388-a0e2-1e75c6e22738","Type":"ContainerStarted","Data":"fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37"} Dec 09 18:43:22 crc kubenswrapper[4954]: I1209 18:43:22.122756 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m49dd" podStartSLOduration=2.600276996 podStartE2EDuration="5.122697996s" podCreationTimestamp="2025-12-09 18:43:17 +0000 UTC" firstStartedPulling="2025-12-09 18:43:19.052507893 +0000 UTC m=+6395.440681723" lastFinishedPulling="2025-12-09 18:43:21.574928903 +0000 UTC m=+6397.963102723" observedRunningTime="2025-12-09 18:43:22.116693708 +0000 UTC m=+6398.504867538" watchObservedRunningTime="2025-12-09 18:43:22.122697996 +0000 UTC m=+6398.510871826" Dec 09 18:43:22 crc kubenswrapper[4954]: E1209 18:43:22.128255 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:43:27 crc kubenswrapper[4954]: E1209 18:43:27.123301 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.389112 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.389156 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.450768 4954 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cpzkt"] Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.453804 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.454166 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.466654 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cpzkt"] Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.489405 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-utilities\") pod \"certified-operators-cpzkt\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.489476 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9pmn\" (UniqueName: \"kubernetes.io/projected/812c0aa3-f831-445f-b08d-7427a705d734-kube-api-access-z9pmn\") pod \"certified-operators-cpzkt\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.489616 4954 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-catalog-content\") pod \"certified-operators-cpzkt\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.591619 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-utilities\") pod \"certified-operators-cpzkt\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.591685 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9pmn\" (UniqueName: \"kubernetes.io/projected/812c0aa3-f831-445f-b08d-7427a705d734-kube-api-access-z9pmn\") pod \"certified-operators-cpzkt\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.591744 4954 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-catalog-content\") pod \"certified-operators-cpzkt\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.592481 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-catalog-content\") pod \"certified-operators-cpzkt\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.592743 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-utilities\") pod \"certified-operators-cpzkt\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.614224 4954 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9pmn\" (UniqueName: \"kubernetes.io/projected/812c0aa3-f831-445f-b08d-7427a705d734-kube-api-access-z9pmn\") pod \"certified-operators-cpzkt\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:27 crc kubenswrapper[4954]: I1209 18:43:27.779874 4954 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:28 crc kubenswrapper[4954]: I1209 18:43:28.232694 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:28 crc kubenswrapper[4954]: I1209 18:43:28.362148 4954 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cpzkt"] Dec 09 18:43:28 crc kubenswrapper[4954]: W1209 18:43:28.367006 4954 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod812c0aa3_f831_445f_b08d_7427a705d734.slice/crio-2013d948e68f92151ce8a8b72f97b7d5cbf2824b303966839d3117854498e011 WatchSource:0}: Error finding container 2013d948e68f92151ce8a8b72f97b7d5cbf2824b303966839d3117854498e011: Status 404 returned error can't find the container with id 2013d948e68f92151ce8a8b72f97b7d5cbf2824b303966839d3117854498e011 Dec 09 18:43:29 crc kubenswrapper[4954]: I1209 18:43:29.180359 4954 generic.go:334] "Generic (PLEG): container finished" podID="812c0aa3-f831-445f-b08d-7427a705d734" containerID="019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90" exitCode=0 Dec 09 18:43:29 crc kubenswrapper[4954]: I1209 18:43:29.180749 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpzkt" event={"ID":"812c0aa3-f831-445f-b08d-7427a705d734","Type":"ContainerDied","Data":"019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90"} Dec 09 18:43:29 crc kubenswrapper[4954]: I1209 18:43:29.180800 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpzkt" event={"ID":"812c0aa3-f831-445f-b08d-7427a705d734","Type":"ContainerStarted","Data":"2013d948e68f92151ce8a8b72f97b7d5cbf2824b303966839d3117854498e011"} Dec 09 18:43:31 crc kubenswrapper[4954]: I1209 18:43:31.201928 4954 generic.go:334] "Generic (PLEG): container finished" podID="812c0aa3-f831-445f-b08d-7427a705d734" containerID="00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3" exitCode=0 Dec 09 18:43:31 crc kubenswrapper[4954]: I1209 18:43:31.202191 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpzkt" event={"ID":"812c0aa3-f831-445f-b08d-7427a705d734","Type":"ContainerDied","Data":"00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3"} Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.213962 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpzkt" event={"ID":"812c0aa3-f831-445f-b08d-7427a705d734","Type":"ContainerStarted","Data":"c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2"} Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.231482 4954 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cpzkt" podStartSLOduration=2.7504275590000002 podStartE2EDuration="5.231459545s" podCreationTimestamp="2025-12-09 18:43:27 +0000 UTC" firstStartedPulling="2025-12-09 18:43:29.183800477 +0000 UTC m=+6405.571974337" lastFinishedPulling="2025-12-09 18:43:31.664832493 +0000 UTC m=+6408.053006323" observedRunningTime="2025-12-09 18:43:32.229171834 +0000 UTC m=+6408.617345674" watchObservedRunningTime="2025-12-09 18:43:32.231459545 +0000 UTC m=+6408.619633375" Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.249232 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m49dd"] Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.249509 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m49dd" podUID="362670c4-832d-4388-a0e2-1e75c6e22738" containerName="registry-server" containerID="cri-o://fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37" gracePeriod=2 Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.761649 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.796169 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57xvd\" (UniqueName: \"kubernetes.io/projected/362670c4-832d-4388-a0e2-1e75c6e22738-kube-api-access-57xvd\") pod \"362670c4-832d-4388-a0e2-1e75c6e22738\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.796415 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-utilities\") pod \"362670c4-832d-4388-a0e2-1e75c6e22738\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.796442 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-catalog-content\") pod \"362670c4-832d-4388-a0e2-1e75c6e22738\" (UID: \"362670c4-832d-4388-a0e2-1e75c6e22738\") " Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.797116 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-utilities" (OuterVolumeSpecName: "utilities") pod "362670c4-832d-4388-a0e2-1e75c6e22738" (UID: "362670c4-832d-4388-a0e2-1e75c6e22738"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.805984 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/362670c4-832d-4388-a0e2-1e75c6e22738-kube-api-access-57xvd" (OuterVolumeSpecName: "kube-api-access-57xvd") pod "362670c4-832d-4388-a0e2-1e75c6e22738" (UID: "362670c4-832d-4388-a0e2-1e75c6e22738"). InnerVolumeSpecName "kube-api-access-57xvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.845383 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "362670c4-832d-4388-a0e2-1e75c6e22738" (UID: "362670c4-832d-4388-a0e2-1e75c6e22738"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.898022 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.898490 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362670c4-832d-4388-a0e2-1e75c6e22738-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:43:32 crc kubenswrapper[4954]: I1209 18:43:32.898588 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57xvd\" (UniqueName: \"kubernetes.io/projected/362670c4-832d-4388-a0e2-1e75c6e22738-kube-api-access-57xvd\") on node \"crc\" DevicePath \"\"" Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.225267 4954 generic.go:334] "Generic (PLEG): container finished" podID="362670c4-832d-4388-a0e2-1e75c6e22738" containerID="fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37" exitCode=0 Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.228023 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m49dd" Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.238345 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m49dd" event={"ID":"362670c4-832d-4388-a0e2-1e75c6e22738","Type":"ContainerDied","Data":"fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37"} Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.238541 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m49dd" event={"ID":"362670c4-832d-4388-a0e2-1e75c6e22738","Type":"ContainerDied","Data":"dd546196a234d32bf72ebe4354dfcfe289e628ea15adae0bb70f4e3eec1fa0ed"} Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.238664 4954 scope.go:117] "RemoveContainer" containerID="fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37" Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.271370 4954 scope.go:117] "RemoveContainer" containerID="36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7" Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.273223 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m49dd"] Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.285141 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m49dd"] Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.296786 4954 scope.go:117] "RemoveContainer" containerID="a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc" Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.344495 4954 scope.go:117] "RemoveContainer" containerID="fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37" Dec 09 18:43:33 crc kubenswrapper[4954]: E1209 18:43:33.345020 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37\": container with ID starting with fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37 not found: ID does not exist" containerID="fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37" Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.345062 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37"} err="failed to get container status \"fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37\": rpc error: code = NotFound desc = could not find container \"fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37\": container with ID starting with fda4382c3176e3be9565fc16480825bdb7a17ef4d715d64f8fce44472cb1cb37 not found: ID does not exist" Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.345090 4954 scope.go:117] "RemoveContainer" containerID="36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7" Dec 09 18:43:33 crc kubenswrapper[4954]: E1209 18:43:33.345414 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7\": container with ID starting with 36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7 not found: ID does not exist" containerID="36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7" Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.345460 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7"} err="failed to get container status \"36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7\": rpc error: code = NotFound desc = could not find container \"36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7\": container with ID starting with 36acb62842628cc6d7fc44f1b81dfc3f3eb37997c2bd07404c937f06558517f7 not found: ID does not exist" Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.345489 4954 scope.go:117] "RemoveContainer" containerID="a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc" Dec 09 18:43:33 crc kubenswrapper[4954]: E1209 18:43:33.345716 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc\": container with ID starting with a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc not found: ID does not exist" containerID="a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc" Dec 09 18:43:33 crc kubenswrapper[4954]: I1209 18:43:33.345741 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc"} err="failed to get container status \"a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc\": rpc error: code = NotFound desc = could not find container \"a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc\": container with ID starting with a8df4c779a76cb84f565525c417f967a6255f58b7ad661d702df776131b6a4fc not found: ID does not exist" Dec 09 18:43:34 crc kubenswrapper[4954]: I1209 18:43:34.132009 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="362670c4-832d-4388-a0e2-1e75c6e22738" path="/var/lib/kubelet/pods/362670c4-832d-4388-a0e2-1e75c6e22738/volumes" Dec 09 18:43:37 crc kubenswrapper[4954]: E1209 18:43:37.124699 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:43:37 crc kubenswrapper[4954]: I1209 18:43:37.780727 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:37 crc kubenswrapper[4954]: I1209 18:43:37.780778 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:37 crc kubenswrapper[4954]: I1209 18:43:37.828936 4954 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:38 crc kubenswrapper[4954]: I1209 18:43:38.318914 4954 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:40 crc kubenswrapper[4954]: I1209 18:43:40.040411 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cpzkt"] Dec 09 18:43:40 crc kubenswrapper[4954]: I1209 18:43:40.291404 4954 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cpzkt" podUID="812c0aa3-f831-445f-b08d-7427a705d734" containerName="registry-server" containerID="cri-o://c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2" gracePeriod=2 Dec 09 18:43:40 crc kubenswrapper[4954]: I1209 18:43:40.799150 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:40 crc kubenswrapper[4954]: I1209 18:43:40.883848 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-catalog-content\") pod \"812c0aa3-f831-445f-b08d-7427a705d734\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " Dec 09 18:43:40 crc kubenswrapper[4954]: I1209 18:43:40.883945 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-utilities\") pod \"812c0aa3-f831-445f-b08d-7427a705d734\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " Dec 09 18:43:40 crc kubenswrapper[4954]: I1209 18:43:40.884200 4954 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9pmn\" (UniqueName: \"kubernetes.io/projected/812c0aa3-f831-445f-b08d-7427a705d734-kube-api-access-z9pmn\") pod \"812c0aa3-f831-445f-b08d-7427a705d734\" (UID: \"812c0aa3-f831-445f-b08d-7427a705d734\") " Dec 09 18:43:40 crc kubenswrapper[4954]: I1209 18:43:40.885182 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-utilities" (OuterVolumeSpecName: "utilities") pod "812c0aa3-f831-445f-b08d-7427a705d734" (UID: "812c0aa3-f831-445f-b08d-7427a705d734"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:43:40 crc kubenswrapper[4954]: I1209 18:43:40.890675 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/812c0aa3-f831-445f-b08d-7427a705d734-kube-api-access-z9pmn" (OuterVolumeSpecName: "kube-api-access-z9pmn") pod "812c0aa3-f831-445f-b08d-7427a705d734" (UID: "812c0aa3-f831-445f-b08d-7427a705d734"). InnerVolumeSpecName "kube-api-access-z9pmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 18:43:40 crc kubenswrapper[4954]: I1209 18:43:40.987573 4954 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 18:43:40 crc kubenswrapper[4954]: I1209 18:43:40.987628 4954 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9pmn\" (UniqueName: \"kubernetes.io/projected/812c0aa3-f831-445f-b08d-7427a705d734-kube-api-access-z9pmn\") on node \"crc\" DevicePath \"\"" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.119125 4954 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "812c0aa3-f831-445f-b08d-7427a705d734" (UID: "812c0aa3-f831-445f-b08d-7427a705d734"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.192207 4954 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/812c0aa3-f831-445f-b08d-7427a705d734-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.303716 4954 generic.go:334] "Generic (PLEG): container finished" podID="812c0aa3-f831-445f-b08d-7427a705d734" containerID="c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2" exitCode=0 Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.303760 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpzkt" event={"ID":"812c0aa3-f831-445f-b08d-7427a705d734","Type":"ContainerDied","Data":"c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2"} Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.303786 4954 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpzkt" event={"ID":"812c0aa3-f831-445f-b08d-7427a705d734","Type":"ContainerDied","Data":"2013d948e68f92151ce8a8b72f97b7d5cbf2824b303966839d3117854498e011"} Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.303801 4954 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cpzkt" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.303802 4954 scope.go:117] "RemoveContainer" containerID="c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.326249 4954 scope.go:117] "RemoveContainer" containerID="00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.349388 4954 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cpzkt"] Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.355586 4954 scope.go:117] "RemoveContainer" containerID="019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.358922 4954 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cpzkt"] Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.406406 4954 scope.go:117] "RemoveContainer" containerID="c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2" Dec 09 18:43:41 crc kubenswrapper[4954]: E1209 18:43:41.406941 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2\": container with ID starting with c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2 not found: ID does not exist" containerID="c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.406997 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2"} err="failed to get container status \"c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2\": rpc error: code = NotFound desc = could not find container \"c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2\": container with ID starting with c10261b360cf63c517ee5e385aa4935af5afce33516522087f9e3e7da2d5bee2 not found: ID does not exist" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.407036 4954 scope.go:117] "RemoveContainer" containerID="00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3" Dec 09 18:43:41 crc kubenswrapper[4954]: E1209 18:43:41.408659 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3\": container with ID starting with 00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3 not found: ID does not exist" containerID="00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.408694 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3"} err="failed to get container status \"00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3\": rpc error: code = NotFound desc = could not find container \"00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3\": container with ID starting with 00fe97291fdc2cd40ccd340bde95be8d92d3f8f4c8aaab99d4c83880c59c5be3 not found: ID does not exist" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.408718 4954 scope.go:117] "RemoveContainer" containerID="019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90" Dec 09 18:43:41 crc kubenswrapper[4954]: E1209 18:43:41.409303 4954 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90\": container with ID starting with 019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90 not found: ID does not exist" containerID="019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90" Dec 09 18:43:41 crc kubenswrapper[4954]: I1209 18:43:41.409347 4954 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90"} err="failed to get container status \"019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90\": rpc error: code = NotFound desc = could not find container \"019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90\": container with ID starting with 019941cb5c0d0db8b28e9eeb62ead618fcfa7fe50557e7be1a5c741b014b6b90 not found: ID does not exist" Dec 09 18:43:42 crc kubenswrapper[4954]: E1209 18:43:42.124463 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38" Dec 09 18:43:42 crc kubenswrapper[4954]: I1209 18:43:42.136366 4954 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="812c0aa3-f831-445f-b08d-7427a705d734" path="/var/lib/kubelet/pods/812c0aa3-f831-445f-b08d-7427a705d734/volumes" Dec 09 18:43:50 crc kubenswrapper[4954]: I1209 18:43:50.122906 4954 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 18:43:50 crc kubenswrapper[4954]: E1209 18:43:50.241023 4954 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:43:50 crc kubenswrapper[4954]: E1209 18:43:50.241114 4954 kuberuntime_image.go:55] "Failed to pull image" err="initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 09 18:43:50 crc kubenswrapper[4954]: E1209 18:43:50.241300 4954 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h8chc5h599h56ch66dh77h5cch646h56ch555h64ch588h596h597h666h76hbbh78h564h664hc8h97h574h589h58h65ch54chcfh595h655h89q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df): ErrImagePull: initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine" logger="UnhandledError" Dec 09 18:43:50 crc kubenswrapper[4954]: E1209 18:43:50.242554 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"initializing source docker://quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested: reading manifest current-tested in quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central: unknown: Tag current-tested was deleted or has expired. To pull, revive via time machine\"" pod="openstack/ceilometer-0" podUID="9efab0ae-1f4e-4dfd-b859-4e3dfe1b28df" Dec 09 18:43:56 crc kubenswrapper[4954]: E1209 18:43:56.125633 4954 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-b2pg9" podUID="8da9d86f-0299-4c30-8ed9-6d6f6ecf4d38"